logo for The Institution of Engineering and Technology
Access Control and Security Monitoring of Multimedia Information Processing and Transmission
Zhihan Lyu
The Institution of Engineering and Technology, 2024
In the era of big data and multi-connectivity via IoTs, protecting and securing multimedia data has become a real necessity and priority for organizations and businesses, but this can be a rather difficult task due to the heterogeneous nature of platforms and data sets. It is therefore essential to improve the security level of multimedia information by developing core technologies to prevent the loss and damage of information during processing and transmission.
[more]

front cover of Advances in Cognitive Systems
Advances in Cognitive Systems
Samia Nefti
The Institution of Engineering and Technology, 2010
This book has been inspired by the portfolio of recent scientific outputs from a range of European and national research initiatives in cognitive science. It presents an overview of recent developments in cognition research and unites the various emerging research strands within a single text as a reference point for progress in the subject. It also provides guidance for new researchers on the breadth of the field and the interconnections between the various research strands identified here.
[more]

logo for The Institution of Engineering and Technology
Affective Computing Applications using Artificial Intelligence in Healthcare
Methods, approaches and challenges in system design
M. Murugappan
The Institution of Engineering and Technology, 2024
Affective computing is the study and development of systems and devices that can recognise human emotions. This can be done using sensing technologies and AI algorithms to process biological signals or facial images to identify the different affective states, such as happiness, anger, fear, surprise, sadness and disgust. This non-invasive technique has applications in healthcare such as emotional impairment detection, mental health assessment, emotional stress assessment, cognitive decline detection, attention deficit disorders, neurodegenerative diseases, neurological disorders, autism spectrum disorder, stress, anxiety or other behavioural assessment.
[more]

front cover of AI for Emerging Verticals
AI for Emerging Verticals
Human-robot computing, sensing and networking
Muhammad Zeeshan Shakir
The Institution of Engineering and Technology, 2021
By specializing in a vertical market, companies can better understand their customers and bring more insight to clients in order to become an integral part of their businesses. This approach requires dedicated tools, which is where artificial intelligence (AI) and machine learning (ML) will play a major role. By adopting AI software and services, businesses can create predictive strategies, enhance their capabilities, better interact with customers, and streamline their business processes.
[more]

front cover of Applications of Artificial Intelligence in E-Healthcare Systems
Applications of Artificial Intelligence in E-Healthcare Systems
Munish Sabharwal
The Institution of Engineering and Technology, 2022
Increased use of artificial intelligence (AI) is being deployed in many hospitals and healthcare settings to help improve health care service delivery. Machine learning (ML) and deep learning (DL) tools can help guide physicians with tasks such as diagnosis and detection of diseases and assisting with medical decision making.
[more]

front cover of Artificial Intelligence and Blockchain Technology in Modern Telehealth Systems
Artificial Intelligence and Blockchain Technology in Modern Telehealth Systems
Agbotiname Lucky Imoize
The Institution of Engineering and Technology, 2024
The expansion of telehealth services is enabling healthcare professionals to consult, diagnose, advise or perform tasks remotely, enabling them to treat more patients in their own homes or consult on cases on the other side of the world. The security of sensitive user information is critical to effective and efficient delivery of healthcare services. Artificial intelligence (AI) and blockchain technology are identified as key drivers of emerging telehealth systems, enabling efficient delivery of telehealth services to billions of patients globally. Specifically, AI facilitates the processing and analysis of complex telehealth data, and blockchain technology offers decentralised, transparent, traceable, reliable, trustful, and provable security to telehealth systems.
[more]

front cover of Artificial Intelligence Applied to Satellite-based Remote Sensing Data for Earth Observation
Artificial Intelligence Applied to Satellite-based Remote Sensing Data for Earth Observation
Maria Pia Del Rosso
The Institution of Engineering and Technology, 2021
Earth observation (EO) involves the collection, analysis, and presentation of data in order to monitor and assess the status and changes in natural and built environments. This technology has many applications including weather forecasting, tracking biodiversity, measuring land-use change, monitoring and responding to natural disasters, managing natural resources, monitoring emerging diseases and health risks, and predicting, adapting to and mitigating climate change.
[more]

front cover of Artificial Intelligence for Biometrics and Cybersecurity
Artificial Intelligence for Biometrics and Cybersecurity
Technology and applications
Ahmed A. Abd El-Latif
The Institution of Engineering and Technology, 2023
The integration of new technologies is resulting in an increased demand for security and authentication in all types of data communications. Cybersecurity is the protection of networks and systems from theft. Biometric technologies use unique traits of particular parts of the body such facial recognition, iris, fingerprints and voice to identify individuals' physical and behavioural characteristics. Although there are many challenges associated with extracting, storing and processing such data, biometric and cybersecurity technologies along with artificial intelligence (AI) are offering new approaches to verification procedures and mitigating security risks.
[more]

front cover of Artificial Intelligence In Libraries And Publishing
Artificial Intelligence In Libraries And Publishing
Ruth Pickering
Against the Grain, LLC, 2022
What is the current state of artificial intelligence (AI) in the world of scholarly communication? What impact does AI have on the practices and strategies of publishers, libraries, information technology companies, and researchers? What exactly is AI and what are those in the realm of scholarly communication actually thinking about it and doing with it?

This Charleston Briefing seeks to provide some answers to these very important questions, offering both general essays on AI and more specific essays on AI in scholarly publishing, academic libraries, and AI in information discovery and knowledge building. The essays will help publishers, librarians, and researchers better understand the actual impact of AI on libraries and publishing so that they can respond to the potentially transformative impact of AI in a measured and knowledgeable manner.

Charleston Briefings: Trending Topics for Information Professionals is a thought-provoking series of brief books concerning innovation in the sphere of libraries, publishing, and technology in scholarly communication. The briefings, growing out of the vital conversations characteristic of the Charleston Conference and Against the Grain, will offer valuable insights into the trends shaping our professional lives and the institutions in which we work.
[more]

front cover of Augmented Exploitation
Augmented Exploitation
Artificial Intelligence, Automation and Work
Phoebe Moore
Pluto Press, 2021
Artificial Intelligence is a seemingly neutral technology, but it is increasingly used to manage workforces and make decisions to hire and fire employees. Its proliferation in the workplace gives the impression of a fairer, more efficient system of management. A machine can't discriminate, after all. Augmented Exploitation explores the reality of the impact of AI on workers' lives. While the consensus is that AI is a completely new way of managing a workplace, the authors show that, on the contrary, AI is used as most technologies are used under capitalism: as a smokescreen that hides the deep exploitation of workers. Going beyond platform work and the gig economy, the authors explore emerging forms of algorithmic governance and AI-augmented apps that have been developed to utilise innovative ways to collect data about workers and consumers, as well as to keep wages and worker representation under control. They also show that workers are not taking this lying down, providing case studies of new and exciting form of resistance that are springing up across the globe.
[more]

front cover of The Birth of Computer Vision
The Birth of Computer Vision
James E. Dobson
University of Minnesota Press, 2023

A revealing genealogy of image-recognition techniques and technologies
 

Today’s most advanced neural networks and sophisticated image-analysis methods come from 1950s and ’60s Cold War culture—and many biases and ways of understanding the world from that era persist along with them. Aerial surveillance and reconnaissance shaped all of the technologies that we now refer to as computer vision, including facial recognition. The Birth of Computer Vision uncovers these histories and finds connections between the algorithms, people, and politics at the core of automating perception today.

James E. Dobson reveals how new forms of computerized surveillance systems, high-tech policing, and automated decision-making systems have become entangled, functioning together as a new technological apparatus of social control. Tracing the development of a series of important computer-vision algorithms, he uncovers the ideas, worrisome military origins, and lingering goals reproduced within the code and the products based on it, examining how they became linked to one another and repurposed for domestic and commercial uses. Dobson includes analysis of the Shakey Project, which produced the first semi-autonomous robot, and the impact of student protest in the early 1970s at Stanford University, as well as recovering the computer vision–related aspects of Frank Rosenblatt’s Perceptron as the crucial link between machine learning and computer vision.

Motivated by the ongoing use of these major algorithms and methods, The Birth of Computer Vision chronicles the foundations of computer vision and artificial intelligence, its major transformations, and the questionable legacy of its origins.


Cover alt text: Two overlapping circles in cream and violet, with black background. Top is a printed circuit with camera eye; below a person at a 1977 computer.

[more]

front cover of Collaborative Language Engineering
Collaborative Language Engineering
A Case Study in Efficient Grammar-Based Processing
Edited by Stephan Oepen, Dan Flickinger, Jun-ichi Tsujii, and Hans Uszkoreit
CSLI, 2001
Following high hopes and subsequent disillusionment in the late 1980s, the past decade of work in language engineering has seen a dramatic increase in the power and sophistication of statistical approaches to natural language processing, along with a growing recognition that these methods alone cannot meet the full range of demands for applications of NLP. While statistical methods, often described as 'shallow' processing techniques, can bring real advantages in robustness and efficiency, they do not provide the precise, reliable representations of meaning which more conventional symbolic grammars can supply for natural language. A consistent, fine-grained mapping between form and meaning is of critical importance in some NLP applications, including machine translation, speech prosthesis, and automated email response. Recent advances in grammar development and processing implementations offer hope of meeting these demands for precision.

This volume provides an update on the state of the art in the development and application of broad-coverage declarative grammars built on sound linguistic foundations - the 'deep' processing paradigm - and presents several aspects of an international research effort to produce comprehensive, re-usable grammars and efficient technology for parsing and generating with such grammars.
[more]

front cover of Collected Papers of Martin Kay
Collected Papers of Martin Kay
A Half Century of Computational Linguistics
Martin Kay, with the editorial assistance of Dan Flickinger and Stephan Oepen
CSLI, 2010
Since the dawn of the age of computers, researchers have been pushing the limits of available processing power to tackle the formidable challenge of developing software that can understand ordinary human language.  At the forefront of this quest for the past fifty years, Martin Kay has been a constant source of new algorithms which have proven fundamental to progress in computational linguistics. Collected Papers of Martin Kay, the first comprehensive collection of his works to date, opens a window into the growth of an increasingly important field of scientific research and development. 
 
 
[more]

front cover of Computer Vision and Recognition Systems Using Machine and Deep Learning Approaches
Computer Vision and Recognition Systems Using Machine and Deep Learning Approaches
Fundamentals, technologies and applications
Chiranji Lal Chowdhary
The Institution of Engineering and Technology, 2021
Computer vision is an interdisciplinary scientific field that deals with how computers obtain, store, interpret and understand digital images or videos using artificial intelligence based on neural networks, machine learning and deep learning methodologies. They are used in countless applications such as image retrieval and classification, driving and transport monitoring, medical diagnostics and aerial monitoring.
[more]

front cover of Deep Learning in Medical Image Processing and Analysis
Deep Learning in Medical Image Processing and Analysis
Khaled Rabie
The Institution of Engineering and Technology, 2023
Medical images, in various formats, are used by clinicians to identify abnormalities or markers associated with certain conditions, such as cancers, diseases, abnormalities or other adverse health conditions. Deep learning algorithms use vast volumes of data to train the computer to recognise certain features in the images that are associated with the disease or condition that you wish to identify.
[more]

logo for University of Pittsburgh Press
The Descent of Artificial Intelligence
A Deep History of an Idea 400 Years in the Making
Kevin Donnelly
University of Pittsburgh Press, 2023
The idea that a new technology could challenge human intelligence is as old as the warning from Socrates and Plato that written language eroded memory. With the emergence of generative artificial intelligence programs, we find ourselves once again debating how a new technology might influence human thought and behavior. Researchers, software developers, and “visionary” tech writers even imagine an AI that will equal or surpass human intelligence, adding to a sense of technological determinism where humanity is inexorably shaped by powerful new machines. But among the hundreds of essays, books, and movies that approach the question of AI, few have asked how exactly scientists and philosophers have codified human thought and behavior. Rather than focusing on technical contributions in machine building, The Descent of Artificial Intelligence explores a more diverse cast of thinkers who helped to imagine the very kind of human being that might be challenged by a machine. Kevin Padraic Donnelly argues that what we often think of as the “goal” of AI has in fact been shaped by forgotten and discredited theories about people and human nature as much as it has been by scientific discoveries, mathematical advances, and novel technologies. By looking at the development of artificial intelligence through the lens of social thought, Donnelly deflates the image of artificial intelligence as a technological monolith and reminds readers that we can control the narratives about ourselves.    
[more]

front cover of Digital Twin Technologies for Healthcare 4.0
Digital Twin Technologies for Healthcare 4.0
Rajesh Kumar Dhanaraj
The Institution of Engineering and Technology, 2022
In healthcare, a digital twin is a digital representation of a patient or healthcare system using integrated simulations and service data. The digital twin tracks a patient's records, crosschecks them against registered patterns and analyses any diseases or contra indications. The digital twin uses adaptive analytics and algorithms to produce accurate prognoses and suggest appropriate interventions. A digital twin can run various medical scenarios before treatment is initiated on the patient, thus increasing patient safety as well as providing the most appropriate treatments to meet the patient's requirements.
[more]

front cover of The Economics of Artificial Intelligence
The Economics of Artificial Intelligence
An Agenda
Edited by Ajay Agrawal, Joshua Gans, and Avi Goldfarb
University of Chicago Press, 2019
Advances in artificial intelligence (AI) highlight the potential of this technology to affect productivity, growth, inequality, market power, innovation, and employment. This volume seeks to set the agenda for economic research on the impact of AI. It covers four broad themes: AI as a general purpose technology; the relationships between AI, growth, jobs, and inequality; regulatory responses to changes brought on by AI; and the effects of AI on the way economic research is conducted. It explores the economic influence of machine learning, the branch of computational statistics that has driven much of the recent excitement around AI, as well as the economic impact of robotics and automation and the potential economic consequences of a still-hypothetical artificial general intelligence. The volume provides frameworks for understanding the economic impact of AI and identifies a number of open research questions.

Contributors:
Daron Acemoglu, Massachusetts Institute of Technology
Philippe Aghion, Collège de France
Ajay Agrawal, University of Toronto
Susan Athey, Stanford University
James Bessen, Boston University School of Law
Erik Brynjolfsson, MIT Sloan School of Management
Colin F. Camerer, California Institute of Technology
Judith Chevalier, Yale School of Management
Iain M. Cockburn, Boston University
Tyler Cowen, George Mason University
Jason Furman, Harvard Kennedy School
Patrick Francois, University of British Columbia 
Alberto Galasso, University of Toronto
Joshua Gans, University of Toronto
Avi Goldfarb, University of Toronto
Austan Goolsbee, University of Chicago Booth School of Business
Rebecca Henderson, Harvard Business School
Ginger Zhe Jin, University of Maryland
Benjamin F. Jones, Northwestern University
Charles I. Jones, Stanford University
Daniel Kahneman, Princeton University
Anton Korinek, Johns Hopkins University
Mara Lederman, University of Toronto
Hong Luo, Harvard Business School
John McHale, National University of Ireland
Paul R. Milgrom, Stanford University
Matthew Mitchell, University of Toronto
Alexander Oettl, Georgia Institute of Technology
Andrea Prat, Columbia Business School
Manav Raj, New York University
Pascual Restrepo, Boston University
Daniel Rock, MIT Sloan School of Management
Jeffrey D. Sachs, Columbia University
Robert Seamans, New York University
Scott Stern, MIT Sloan School of Management
Betsey Stevenson, University of Michigan
Joseph E. Stiglitz. Columbia University
Chad Syverson, University of Chicago Booth School of Business
Matt Taddy, University of Chicago Booth School of Business
Steven Tadelis, University of California, Berkeley
Manuel Trajtenberg, Tel Aviv University
Daniel Trefler, University of Toronto
Catherine Tucker, MIT Sloan School of Management
Hal Varian, University of California, Berkeley
[more]

front cover of The Economics of Artificial Intelligence
The Economics of Artificial Intelligence
Health Care Challenges
Edited by Ajay Agrawal, Joshua Gans, Avi Goldfarb, and Catherine E. Tucker
University of Chicago Press, 2024

A timely investigation of the potential economic effects, both realized and unrealized, of artificial intelligence within the United States healthcare system.

In sweeping conversations about the impact of artificial intelligence on many sectors of the economy, healthcare has received relatively little attention. Yet it seems unlikely that an industry that represents nearly one-fifth of the economy could escape the efficiency and cost-driven disruptions of AI.

The Economics of Artificial Intelligence: Health Care Challenges brings together contributions from health economists, physicians, philosophers, and scholars in law, public health, and machine learning to identify the primary barriers to entry of AI in the healthcare sector. Across original papers and in wide-ranging responses, the contributors analyze barriers of four types: incentives, management, data availability, and regulation. They also suggest that AI has the potential to improve outcomes and lower costs. Understanding both the benefits of and barriers to AI adoption is essential for designing policies that will affect the evolution of the healthcare system.

[more]

front cover of Evolving Predictive Analytics in Healthcare
Evolving Predictive Analytics in Healthcare
New AI techniques for real-time interventions
Abhishek Kumar
The Institution of Engineering and Technology, 2022
A major use of practical predictive analytics in medicine has been in the diagnosis of current diseases, particularly through medical imaging. Now there is sufficient improvement in AI, IoT and data analytics to deal with real time problems with an increased focus on early prediction using machine learning and deep learning algorithms. With the power of artificial intelligence alongside the internet of 'medical' things, these algorithms can input the characteristics/data of their patients and get predictions of future diagnoses, classifications, treatment and costs.
[more]

front cover of Experimenting the Human
Experimenting the Human
Art, Music, and the Contemporary Posthuman
G Douglas Barrett
University of Chicago Press, 2023
An engaging argument about what experimental music can tell us about being human.

In Experimenting the Human, G Douglas Barrett argues that experimental music speaks to the contemporary posthuman, a condition in which science and technology decenter human agency amid the uneven temporality of postwar global capitalism. Time moves forward for some during this period, while it seems to stand still or even move backward for others. Some say we’re already posthuman, while others endure the extended consequences of never having been considered fully human in the first place. Experimental music reflects on this state, Barrett contends, through its interdisciplinary involvements in postwar science, technology, and art movements.

Rather than pursuing the human's beyond, experimental music addresses the social and technological conditions that support such a pursuit. Barrett locates this tendency of experimentalism throughout its historical entanglements with cybernetics, and in his intimate analysis of Alvin Lucier’s neurofeedback music, Pamela Z’s BodySynth performances, Nam June Paik’s musical robotics, Pauline Oliveros’s experiments with radio astronomy, and work by Laetitia Sonami, Yasunao Tone, and Jerry Hunt. Through a unique meeting of music studies, media theory, and art history, Experimenting the Human provides fresh insights into what it means to be human.

 
[more]

logo for The Institution of Engineering and Technology
Explainable Artificial Intelligence for Trustworthy Internet of Things
Mohamed Abdel-Basset
The Institution of Engineering and Technology, 2024
A major challenge for machine learning solutions is that their efficiency in real-world applications is constrained by the current lack of ability of the machine to explain its decisions and activities to human users. Biases based on race, gender, age or location have been a long-standing risk in training AI models. Furthermore, AI model performance can degrade because production data differs from training data.
[more]

front cover of Explainable Artificial Intelligence (XAI)
Explainable Artificial Intelligence (XAI)
Concepts, enabling tools, technologies and applications
Pethuru Raj
The Institution of Engineering and Technology, 2023
The world is keen to leverage multi-faceted AI techniques and tools to deploy and deliver the next generation of business and IT applications. Resource-intensive gadgets, machines, instruments, appliances, and equipment spread across a variety of environments are empowered with AI competencies. Connected products are collectively or individually enabled to be intelligent in their operations, offering and output.
[more]

front cover of Genesis Redux
Genesis Redux
Essays in the History and Philosophy of Artificial Life
Edited by Jessica Riskin
University of Chicago Press, 2007

Since antiquity, philosophers and engineers have tried to take life’s measure by reproducing it. Aiming to reenact Creation, at least in part, these experimenters have hoped to understand the links between body and spirit, matter and mind, mechanism and consciousness. Genesis Redux examines moments from this centuries-long experimental tradition: efforts to simulate life in machinery, to synthesize life out of material parts, and to understand living beings by comparison with inanimate mechanisms.

Jessica Riskin collects seventeen essays from distinguished scholars in several fields. These studies offer an unexpected and far-reaching result: attempts to create artificial life have rarely been driven by an impulse to reduce life and mind to machinery.  On the contrary, designers of synthetic creatures have generally assumed a role for something nonmechanical. The history of artificial life is thus also a history of theories of soul and intellect.

Taking a historical approach to a modern quandary, Genesis Redux is essential reading for historians and philosophers of science and technology, scientists and engineers working in artificial life and intelligence, and anyone engaged in evaluating these world-changing projects.

[more]

front cover of Holographic Reduced Representation
Holographic Reduced Representation
Distributed Representation for Cognitive Structures
Tony A. Plate
CSLI, 2003
While neuroscientists garner success in identifying brain regions and in analyzing individual neurons, ground is still being broken at the intermediate scale of understanding how neurons combine to encode information. This book proposes a method of representing information in a computer that would be suited for modeling the brain's methods of processing information.

Holographic Reduced Representations (HRRs) are introduced here to model how the brain distributes each piece of information among thousands of neurons. It had been previously thought that the grammatical structure of a language cannot be encoded practically in a distributed representation, but HRRs can overcome the problems of earlier proposals. Thus this work has implications for psychology, neuroscience, linguistics, and computer science, and engineering.
[more]

front cover of Imaging and Sensing for Unmanned Aircraft Systems
Imaging and Sensing for Unmanned Aircraft Systems
Control and Performance, Volume 1
Vania V. Estrela
The Institution of Engineering and Technology, 2020
This two-volume book set explores how sensors and computer vision technologies are used for the navigation, control, stability, reliability, guidance, fault detection, self-maintenance, strategic re-planning and reconfiguration of unmanned aircraft systems (UAS).
[more]

front cover of Imaging and Sensing for Unmanned Aircraft Systems
Imaging and Sensing for Unmanned Aircraft Systems
Deployment and Applications, Volume 2
Vania V. Estrela
The Institution of Engineering and Technology, 2020
This two-volume book set explores how sensors and computer vision technologies are used for the navigation, control, stability, reliability, guidance, fault detection, self-maintenance, strategic re-planning and reconfiguration of unmanned aircraft systems (UAS).
[more]

logo for Georgetown University Press
Information in War
Military Innovation, Battle Networks, and the Future of Artificial Intelligence
Benjamin M. Jensen, Christopher Whyte, and Scott Cuomo
Georgetown University Press, 2022

An in-depth assessment of innovations in military information technology informs hypothetical outcomes for artificial intelligence adaptations

In the coming decades, artificial intelligence (AI) could revolutionize the way humans wage war. The military organizations that best innovate and adapt to this AI revolution will likely gain significant advantages over their rivals. To this end, great powers such as the United States, China, and Russia are already investing in novel sensing, reasoning, and learning technologies that will alter how militaries plan and fight. The resulting transformation could fundamentally change the character of war.

In Information in War, Benjamin Jensen, Christopher Whyte, and Scott Cuomo provide a deeper understanding of the AI revolution by exploring the relationship between information, organizational dynamics, and military power. The authors analyze how militaries adjust to new information communication technology historically to identify opportunities, risks, and obstacles that will almost certainly confront modern defense organizations as they pursue AI pathways to the future. Information in War builds on these historical cases to frame four alternative future scenarios exploring what the AI revolution could look like in the US military by 2040.

[more]

front cover of Inhuman Power
Inhuman Power
Artificial Intelligence and the Future of Capitalism
Nick Dyer-Witheford, Atle Mikkola Kjosen, and James Steinhoff
Pluto Press, 2019
The past several years have brought staggering advances in the field of Artificial Intelligence. And Marxist analysis has to keep up: while machines were always central to Marxist analysis, modern AI is a new kind of machine that Marx could not have anticipated.
           
Inhuman Power explores the relationship between Marxist theory and AI through three approaches, each using the lens of a different Marxist theoretical concept. While the idea of widespread AI tends to be celebrated as much as questioned, a deeper analysis of its reach and potential produces a more complex and disturbing picture than has been identified. Inhuman Power argues that on its current trajectory, AI is likely to render humanity obsolete and that the only way to prevent it is a communist revolution.
 
[more]

front cover of The Inspiration Machine
The Inspiration Machine
Computational Creativity in Poetry and Jazz
Eitan Y. Wilf
University of Chicago Press, 2023
Explores how creative digital technologies and artificial intelligence are embedded in culture and society.
 
In The Inspiration Machine, Eitan Y. Wilf explores the transformative potentials that digital technology opens up for creative practice through three ethnographic cases, two with jazz musicians and one with a group of poets. At times dissatisfied with the limitations of human creativity, these artists do not turn to computerized algorithms merely to execute their preconceived ideas. Rather, they approach them as creative partners, delegating to them different degrees of agentive control and artistic decision-making in the hopes of finding inspiration in their output and thereby expanding their own creative horizons.
 
The algorithms these artists develop and use, however, remain rooted in and haunted by the specific social predicaments and human shortfalls that they were intended to overcome. Experiments in the digital thus hold an important lesson: although Wilf’s interlocutors returned from their adventures with computational creativity with modified, novel, and enriched capacities and predilections, they also gained a renewed appreciation for, and at times a desire to re-inhabit, non-digital creativity. In examining the potentials and pitfalls of seemingly autonomous digital technologies in the realm of art, Wilf shows that computational solutions to the real or imagined insufficiencies of human practice are best developed in relation to, rather than away from, the social and cultural contexts that gave rise to those insufficiencies, in the first place.
[more]

front cover of Intelligent Distributed Video Surveillance Systems
Intelligent Distributed Video Surveillance Systems
Sergio A. Velastin
The Institution of Engineering and Technology, 2006
There is a growing interest in the development and deployment of surveillance systems in public and private locations. Conventional approaches rely on the installation of wide area CCTV (Closed Circuit Television), but the explosion in the numbers of cameras that have to be monitored, the increasing costs of providing monitoring personnel and the limitations that humans have to maintain sustained levels of concentration severely limit the effectiveness of these systems. Advances in information and communication technologies, such as computer vision for face recognition and human behaviour analysis, digital annotation and storage of video, transmission of video/audio streams over wired and wireless networks, can potentially provide significant improvements in this field.
[more]

front cover of Intelligent Multimedia Processing and Computer Vision
Intelligent Multimedia Processing and Computer Vision
Techniques and applications
Shyam Singh Rajput
The Institution of Engineering and Technology, 2023
Intelligent multimedia involves the computer processing and understanding of perceptual input from speech, text, videos and images. Reacting to these inputs is complex and involves research from engineering, computer science and cognitive science. Intelligent multimedia processing deals with the analysis of images and videos to extract useful information for numerous applications including medical imaging, robotics, remote sensing, autonomous driving, AR/VR, law enforcement, biometrics, multimedia enhancement and reconstruction, agriculture, and security. Intelligent multimedia processing and computer vision have seen an upsurge over the last few years. With the increasing use of intelligent multimedia processing techniques in various sectors, the requirement for fast and reliable techniques to analyse and process multimedia content is increasing day by day.
[more]

front cover of Journalism Ethics
Journalism Ethics
21 Essentials from Wars to Artificial Intelligence
Eric Wishart
Hong Kong University Press, 2024
A necessary guide to responsible journalism in a challenging media landscape.

This concise and authoritative work offers the latest guidance on journalism ethics for students and media professionals and will help empower news consumers to make informed decisions about the trustworthiness of their sources of information. It offers advice on all aspects of journalism ethics including accuracy and seeking the truth, representation of women, LGBTQ coverage, climate change, mental health, use of images, conflict reporting, elections, and how to use artificial intelligence. The author brings a unique perspective and depth of knowledge to the complex challenges facing journalists and news consumers in this era of fake news, disinformation, and artificial intelligence.
[more]

logo for The Institution of Engineering and Technology
Knowledge-based Systems for Industrial Control
J. McGhee
The Institution of Engineering and Technology, 1990
Expert and knowledge-based systems have great potential for industrial control systems, particularly in the process industries. Recognising the importance of this emerging area, the Institution of Electrical Engineers organised a Vacation School on the subject, for engineers from industry and academia, at the University of Strathclyde in September 1990.
[more]

front cover of Language and Learning for Robots
Language and Learning for Robots
Colleen Crangle and Patrick Suppes
CSLI, 1994
Robot technology will find wide-scale use only when a robotic device can be given commands and taught new tasks in a natural language. How could a robot understand instructions expressed in English? How could a robot learn from instructions? Crangle and Suppes begin to answer these questions through a theoretical approach to language and learning for robots and by experimental work with robots.

The authors develop the notion of an instructable robot—one which derives its intelligence in part from interaction with humans. Since verbal interaction with a robot requires a natural language semantics, the authors propose a natural-model semantics which they then apply to the interpretation of robot commands. Two experimental projects are described which provide natural-language interfaces to robotic aids for the physically disabled. The authors discuss the specific challenges posed by the interpretation of "stop" commands and the interpretation of spatial prepositions.

The authors also examine the use of explicit verbal instruction to teach a robot new procedures, propose ways a robot can learn from corrective commands containing qualitative spatial expressions, and discuss the machine-learning of a natural language use to instruct a robot in the performance of simple physical tasks. Two chapters focus on probabilistic techniques in learning.
[more]

front cover of Language and the Rise of the Algorithm
Language and the Rise of the Algorithm
Jeffrey M. Binder
University of Chicago Press, 2022
A wide-ranging history of the algorithm.

Bringing together the histories of mathematics, computer science, and linguistic thought, Language and the Rise of the Algorithm reveals how recent developments in artificial intelligence are reopening an issue that troubled mathematicians well before the computer age: How do you draw the line between computational rules and the complexities of making systems comprehensible to people? By attending to this question, we come to see that the modern idea of the algorithm is implicated in a long history of attempts to maintain a disciplinary boundary separating technical knowledge from the languages people speak day to day.
 
Here Jeffrey M. Binder offers a compelling tour of four visions of universal computation that addressed this issue in very different ways: G. W. Leibniz’s calculus ratiocinator; a universal algebra scheme Nicolas de Condorcet designed during the French Revolution; George Boole’s nineteenth-century logic system; and the early programming language ALGOL, short for algorithmic language. These episodes show that symbolic computation has repeatedly become entangled in debates about the nature of communication. Machine learning, in its increasing dependence on words, erodes the line between technical and everyday language, revealing the urgent stakes underlying this boundary.
 
The idea of the algorithm is a levee holding back the social complexity of language, and it is about to break. This book is about the flood that inspired its construction.
[more]

front cover of Machine Learning for Healthcare Technologies
Machine Learning for Healthcare Technologies
David A. Clifton
The Institution of Engineering and Technology, 2017
This book provides a snapshot of the state of current research at the interface between machine learning and healthcare with special emphasis on machine learning projects that are (or are close to) achieving improvement in patient outcomes. The book provides overviews on a range of technologies including detecting artefactual events in vital signs monitoring data; patient physiological monitoring; tracking infectious disease; predicting antibiotic resistance from genomic data; and managing chronic disease.
[more]

front cover of Machine Learning in Medical Imaging and Computer Vision
Machine Learning in Medical Imaging and Computer Vision
Amita Nandal
The Institution of Engineering and Technology, 2024
Medical images can highlight differences between healthy tissue and unhealthy tissue and these images can then be assessed by a healthcare professional to identify the stage and spread of a disease so a treatment path can be established. With machine learning techniques becoming more prevalent in healthcare, algorithms can be trained to identify healthy or unhealthy tissues and quickly differentiate between the two. Statistical models can be used to process numerous images of the same type in a fraction of the time it would take a human to assess the same quantity, saving time and money in aiding practitioners in their assessment.
[more]

front cover of The Myth of Artificial Intelligence
The Myth of Artificial Intelligence
Why Computers Can’t Think the Way We Do
Erik J. Larson
Harvard University Press, 2021

“Exposes the vast gap between the actual science underlying AI and the dramatic claims being made for it.”
—John Horgan


“If you want to know about AI, read this book…It shows how a supposedly futuristic reverence for Artificial Intelligence retards progress when it denigrates our most irreplaceable resource for any future progress: our own human intelligence.”
—Peter Thiel

Ever since Alan Turing, AI enthusiasts have equated artificial intelligence with human intelligence. A computer scientist working at the forefront of natural language processing, Erik Larson takes us on a tour of the landscape of AI to reveal why this is a profound mistake.

AI works on inductive reasoning, crunching data sets to predict outcomes. But humans don’t correlate data sets. We make conjectures, informed by context and experience. And we haven’t a clue how to program that kind of intuitive reasoning, which lies at the heart of common sense. Futurists insist AI will soon eclipse the capacities of the most gifted mind, but Larson shows how far we are from superintelligence—and what it would take to get there.

“Larson worries that we’re making two mistakes at once, defining human intelligence down while overestimating what AI is likely to achieve…Another concern is learned passivity: our tendency to assume that AI will solve problems and our failure, as a result, to cultivate human ingenuity.”
—David A. Shaywitz, Wall Street Journal

“A convincing case that artificial general intelligence—machine-based intelligence that matches our own—is beyond the capacity of algorithmic machine learning because there is a mismatch between how humans and machines know what they know.”
—Sue Halpern, New York Review of Books

[more]

front cover of Neural Networks
Neural Networks
Ranjodh Singh Dhaliwal
University of Minnesota Press, 2024

A critical examination of the figure of the neural network as it mediates neuroscientific and computational discourses and technical practices

Neural Networks proposes to reconstruct situated practices, social histories, mediating techniques, and ontological assumptions that inform the computational project of the same name. If so-called machine learning comprises a statistical approach to pattern extraction, then neural networks can be defined as a biologically inspired model that relies on probabilistically weighted neuron-like units to identify such patterns. Far from signaling the ultimate convergence of human and machine intelligence, however, neural networks highlight the technologization of neurophysiology that characterizes virtually all strands of neuroscientific and AI research of the past century. Taking this traffic as its starting point, this volume explores how cognition came to be constructed as essentially computational in nature, to the point of underwriting a technologized view of human biology, psychology, and sociability, and how countermovements provide resources for thinking otherwise.

[more]

front cover of Partiality, Modality and Nonmonotonicity
Partiality, Modality and Nonmonotonicity
Patrick Doherty
CSLI, 1996
This edited volume of articles provides a state-of-the-art description of research in logic-based approaches to knowledge representation which combines approaches to reasoning with incomplete information that include partial, modal, and nonmonotonic logics. The collection contains two parts: foundations and case studies. The foundations section provides a general overview of partiality, multi-valued logics, use of modal logic to model partiality and resource-limited inference, and an integration of partial and modal logics. The case studies section provides specific studies of issues raised in the foundations section. Several of the case studies integrate modal and partial modal logics with nonmonotonic logics. Both theoretical and practical aspects of such integration are considered. Knowledge representation issues such as default reasoning, theories of action and change, reason maintenance, awareness, and automation of nonmonotonic reasoning are covered.
[more]

front cover of Passwords
Passwords
Philology, Security, Authentication
Brian Lennon
Harvard University Press, 2018

Cryptology, the mathematical and technical science of ciphers and codes, and philology, the humanistic study of natural or human languages, are typically understood as separate domains of activity. But Brian Lennon contends that these two domains, both concerned with authentication of text, should be viewed as contiguous. He argues that computing’s humanistic applications are as historically important as its mathematical and technical ones. What is more, these humanistic uses, no less than cryptological ones, are marked and constrained by the priorities of security and military institutions devoted to fighting wars and decoding intelligence.

Lennon’s history encompasses the first documented techniques for the statistical analysis of text, early experiments in mechanized literary analysis, electromechanical and electronic code-breaking and machine translation, early literary data processing, the computational philology of late twentieth-century humanities computing, and early twenty-first-century digital humanities. Throughout, Passwords makes clear the continuity between cryptology and philology, showing how the same practices flourish in literary study and in conditions of war.

Lennon emphasizes the convergence of cryptology and philology in the modern digital password. Like philologists, hackers use computational methods to break open the secrets coded in text. One of their preferred tools is the dictionary, that preeminent product of the philologist’s scholarly labor, which supplies the raw material for computational processing of natural language. Thus does the historic overlap of cryptology and philology persist in an artifact of computing—passwords—that many of us use every day.

[more]

front cover of Personal Knowledge Graphs (PKGs)
Personal Knowledge Graphs (PKGs)
Methodology, tools and applications
Sanju Tiwari
The Institution of Engineering and Technology, 2023
Since the development of the semantic web, knowledge graphs (KGs) have been used by search engines, knowledge-engines and question-answering services as well as social networks. A knowledge graph, also known as a semantic network, represents and illustrates a network of real-world entities such as objects, events, situations, or concepts and the relationships between them. This information is usually stored in a graph database and visualized as a graph structure, prompting the term "knowledge graph". Knowledge graphs structure the information of entities, their properties and the relation between them.
[more]

front cover of Perspectives on Contexts
Perspectives on Contexts
Edited by Paolo Bouquet, Luciano Serafini, and Richmond H. Thomason
CSLI, 2008

Most human thinking is thoroughly informed by context but, until recently, theories of reasoning have concentrated on abstract rules and generalities that make no reference to this crucial factor. Perspectives on Contexts brings together essays from leading cognitive scientists to forge a vigorous interdisciplinary understanding of the contextual phenomenon. Applicable to human and machine cognition in philosophy, artificial intelligence, and psychology, this volume is essential to the current renaissance in thinking about context.

[more]

front cover of Putting Linguistics into Speech Recognition
Putting Linguistics into Speech Recognition
The Regulus Grammar Compiler
Manny Rayner, Beth Ann Hockey, and Pierrette Bouillon
CSLI, 2006
Most computer programs that analyze spoken dialogue use a spoken command grammar, which limits what the user can say when talking to the system. To make this process simpler, more automated, and effective for command grammars even at initial stages of a project, the Regulus grammar compiler was developed by a consortium of experts—including NASA scientists. This book presents a complete description of both the practical and theoretical aspects of Regulus and will be extremely helpful for students and scholars working in computational linguistics as well as software engineering.
[more]

front cover of ReRAM-based Machine Learning
ReRAM-based Machine Learning
Hao Yu
The Institution of Engineering and Technology, 2021
The transition towards exascale computing has resulted in major transformations in computing paradigms. The need to analyze and respond to such large amounts of data sets has led to the adoption of machine learning (ML) and deep learning (DL) methods in a wide range of applications.
[more]

front cover of Seeming Human
Seeming Human
Artificial Intelligence and Victorian Realist Character
Megan Ward
The Ohio State University Press, 2018
Seeming Human: Artificial Intelligence and Victorian Realist Character offers a new theory of realist character through character’s unexpected afterlife: the intelligent machine. The book contends that mid-twentieth-century versions of artificial intelligence (AI) offer a theory of verisimilitude omitted by traditional histories of character, which often focus on the development of interiority and the shift from “flat” to “round” characters in the Victorian era. Instead, by reading character through AI, Megan Ward’s Seeming Human argues that routinization, predictability, automation, and even flatness are all features of realist characters.
 
Early artificial intelligence movements such as cybernetics, information theory, and the Turing test define ways of seeming—rather than being—human. Using these theories of verisimilitude to read Victorian novelists such as Elizabeth Gaskell, Margaret Oliphant, Anthony Trollope, Thomas Hardy, and Henry James, Seeming Human argues that mechanicity has been perceived as anti-realist because it is the element that we least want to identify as human. Because AI produces human-like intelligence, it makes clear that we must actually turn to machines in order to understand what makes realist characters seem so human.
 
[more]

front cover of Signal Processing and Machine Learning for Brain-Machine Interfaces
Signal Processing and Machine Learning for Brain-Machine Interfaces
Toshihisa Tanaka
The Institution of Engineering and Technology, 2018
Brain-machine interfacing or brain-computer interfacing (BMI/BCI) is an emerging and challenging technology used in engineering and neuroscience. The ultimate goal is to provide a pathway from the brain to the external world via mapping, assisting, augmenting or repairing human cognitive or sensory-motor functions.
[more]

logo for Georgetown University Press
Strategy, Evolution, and War
From Apes to Artificial Intelligence
Kenneth Payne
Georgetown University Press, 2018

Decisions about war have always been made by humans, but now intelligent machines are on the cusp of changing things – with dramatic consequences for international affairs. This book explores the evolutionary origins of human strategy, and makes a provocative argument that Artificial Intelligence will radically transform the nature of war by changing the psychological basis of decision-making about violence.

Strategy, Evolution, and War is a cautionary preview of how Artificial Intelligence (AI) will revolutionize strategy more than any development in the last three thousand years of military history. Kenneth Payne describes strategy as an evolved package of conscious and unconscious behaviors with roots in our primate ancestry. Our minds were shaped by the need to think about warfare—a constant threat for early humans. As a result, we developed a sophisticated and strategic intelligence.

The implications of AI are profound because they depart radically from the biological basis of human intelligence. Rather than being just another tool of war, AI will dramatically speed up decision making and use very different cognitive processes, including when deciding to launch an attack, or escalate violence. AI will change the essence of strategy, the organization of armed forces, and the international order.

This book is a fascinating examination of the psychology of strategy-making from prehistoric times, through the ancient world, and into the modern age.

[more]

front cover of The Tbilisi Symposium on Logic, Language and Computation
The Tbilisi Symposium on Logic, Language and Computation
Selected Papers
Edited by Jonathan Ginzburg, Zurab Khasidashvili, Carl Vogel, Jean-Jacques Lévy,
CSLI, 1998
This volume brings together papers from linguists, logicians, and computer scientists from thirteen countries (Armenia, Denmark, France, Georgia, Germany, Israel, Italy, Japan, Poland, Spain, Sweden, UK, and USA). This collection aims to serve as a catalyst for new interdisciplinary developments in language, logic and computation and to introduce new ideas from the expanded European academic community. Spanning a wide range of disciplines, the papers cover such topics as formal semantics of natural language, dynamic semantics, channel theory, formal syntax of natural language, formal language theory, corpus-based methods in computational linguistics, computational semantics, syntactic and semantic aspects of l-calculus, non-classical logics, and a fundamental problem in predicate logic.
[more]

front cover of Technologies for Healthcare 4.0
Technologies for Healthcare 4.0
From AI and IoT to blockchain
Karthik Ramamurthy
The Institution of Engineering and Technology, 2023
There are a growing number of challenges in handling medical data in order to provide an effective healthcare service in real-time. Bridging the gap between patient expectations and their experiences needs effective collaboration and connectivity across the healthcare ecosystem. The success of joined-up care relies on patient data being shared between all active stakeholders, including hospitals, outreach workers, and GPs. All these needs and challenges pave the way for the next trend of development in healthcare - healthcare 4.0.
[more]

front cover of Towards a Data-Driven Military
Towards a Data-Driven Military
A Multidisciplinary Perspective
Peter B.M.J. Pijpers
Leiden University Press, 2023
Towards a Data-Driven Military: A Multidisciplinary Perspective assesses the use of data and information on modern conflict from different scientific and methodological disciplines, aiming to generate valuable contributions to the ongoing discourse on data, the military and modern warfare. Part one, ‘Military Systems and Technology’,, approaches the theme empirically by researching how data can enhance the utility of military materiel and subsequently accelerate the decision-making process. Part two, ‘War Studies’, takes a multidisciplinary approach to the evolution of warfare, while the third part, ‘Military Management Studies’, takes a holistic organisational and procedural approach. Based on their scientific protocols and research methods, the three domains put forward different research questions and perspectives, providing the unique character of this book.
[more]

logo for University of Chicago Press
Waiting for Robots
The Hired Hands of Automation
Antonio A. Casilli
University of Chicago Press
An essential investigation that reveals the labor of human workers hidden behind a curtain of apparent technological automation.
 
Artificial Intelligence fuels both enthusiasm and panic. Technologists are inclined to give their creations leeway, pretend they’re animated beings, and consider them efficient. As users, we may complain when these technologies don’t obey, or worry about their influence on our choices and our livelihoods. And yet, we also yearn for their convenience, see ourselves reflected in them, and treat them as something entirely new. But when we overestimate the automation of these tools, award-winning author Antonio A. Casilli argues, we fail to recognize how our fellow humans are essential to their efficiency. The danger is not that robots will take our jobs, but that humans will have to do theirs.
 
In this bracing and powerful book, Casilli uses up-to-the-minute research to show how today’s technologies, including AI, continue to exploit human labor—even ours. He connects the diverse activities of today’s tech laborers: platform workers, like Uber drivers and Airbnb hosts; “micro workers,” including those performing atomized tasks like data entry on Amazon Mechanical Turk; and the rest of us, as we evaluate text or images to show we’re not robots, react to Facebook posts, or approve or improve the output of generative AI. As Casilli shows us, algorithms, search engines, and voice assistants wouldn’t function without unpaid or underpaid human contributions. Further, he warns that if we fail to recognize this human work, we risk a dark future for all human labor.
 
Waiting for Robots urges us to move beyond the simplistic notion that machines are intelligent and autonomous. As the proverbial Godot, robots are the bearers of a messianic promise that is always postponed. Instead of bringing prosperity for all, they discipline the workforce, so we don’t dream of a world without drudgery and exploitation. Casilli’s eye-opening book makes clear that most “automation” requires human labor—and likely always will—shedding new light on today’s consequences and tomorrow’s threats of failing to recognize and compensate the “click workers” of today.
[more]


Send via email Share on Facebook Share on Twitter