THE MEANING OF HUMAN-COMPUTER INTERACTION
An experientially based theoretical HCI-perspective
by Kristo Ivanov, prof.em., Umeå
University
September 2020 (rev. 251115-1645)
(https://archive.org/details/hci_20200910)
(http://www8.informatik.umu.se/~kivanov/HCI.html)
CONTENTS
Link
to a General Disclaimer
Abstract
Introduction and an example
Example
#1: - Health care center
Example
#2 - Telecommunications company
Example
#3 - Telephone answering machines
Example
#4 - Personal identification: PIN-problem
Example
#5 - Public transports and parking
Example
#6 - Bank safe deposit box
Example
#7 - Dental hygienists and antibiotics
Example
#8.- Radio Sweden’s summer program
Example
#9 - Airliners’ and traffic’s control
Quagmire or
intellectual jungle
This is a report on reflections upon
experiences of on-line computer interactions, relating them to insights into
the core meaning of computers, computation and computerization as advanced
mainly in earlier texts on Computers as embodied
mathematics and logic, and Computerization as logic
acrobatics. The
conclusion is that perceived advances in the HCI-field tend to be sterile,
socially and politically crippled patchwork because they do not acknowledge the
nature of the relation between embodied logic as a human mental construct, and
mental faculties of the human psyche.
https://archive.org/details/kant-gramont
This introduction and the following section on
“Further examples” consists mainly of examples of
events that I have witnessed (no sampling except for my memory) and have
personal knowledge of. Most of them are apparently simple or too simple but
with the advantage of being recognizable commonplace. The last example #9 is too complex and exclusive, but
very realistic and crucial. The point is to show that the simplicity hides
complexity. The risk is that apparent simplicity invites to narrow technical
conceptualization of HCI and to simple injunctions to users, whoever they are –
if properly defined - about the use of HCI. This corresponds to patchworking
the shortcomings of logic embodied in computer software with a transfer of
responsibility to users, and it illustrates the meaning of what I elsewhere
called “computerization as logic acrobatics.”
Example
#1 - Health care center
In September 2020 during the covid-19 pandemics, an
eighty years old lady who had an appointment with a doctor at a local Swedish
healthcare center felt indisposed and not able to walk and take the bus to the
center. Not being possible to phone the doctor directly she had to call the
center's computerized telephone exchange. After the call's required initial
selections among a series of alternatives codes, she chose the one for being
called later by a nurse on duty, with an automatically computed and communicated
estimated waiting time of one hour and a half. When the nurse called, the lady
informed her about the by then missed appointment
Two weeks later the old lady got an invoice from the
administrative authorities, corresponding to the double amount that would have
been charged for a common medical consultation, despite of her having been
entitled to a visit free of charge in view of prior accumulated expenses having
reached a maximum according to the Swedish welfare rules. The invoice appeared
to have been produced automatically by a sort of "elementary AI" on
the basis of the computerized data base terminal at the healthcare center and the
missed registration at the expected arrival of the lady at the time of the
appointment, combined with her PIN (personal identification number) linked to
her official home address. It is also the address coupled to the all Swedish
governmental institutions and other entities, including the Swedish Tax Agency and all possible consequent debt collection.
Surprised by the unexpected invoice the lady was
directed to Internet-links (example in Swedish here)
stating that new valid operational rules for health services
"anonymously-legally" stipulated that if an appointment is not
canceled at least 24 hours in advance it will be charged for the stipulated
amount, even if the patient was entitled to free care, is below age 18 or above
age 85. And this is also valid (which is not stated in all relevant links) if
the patient arrives too late to the appointment, when more than half of the
(often but not always unstated twenty minutes or half an hour) allotted time
has passed. Inquiry with the nurse on duty resulted in her response being that
patients all too often simply miss appointments and that they motivate this by
just giving excuses and lies. Patients are simply not deserving to be trusted.
It is interesting to note something that usually only appears in an “insändare” ["letters to the editor"] of small
local newspapers such as Mitti Värmdö (15
September 2020) with the title "Läkaren är alltid sen" [The doctor
is always late]: a patient reports, against the background of having to wait
5-6 weeks for an appointment lasting 30 minutes, that the doctor often appears
after a delay of about 20 minutes, sometimes up to one hour. Not to mention waiting
times of up to 12 hours at an emergency room of a clinic or hospital.
The lady in question wondered about her own prior
experiences of having to wait at the healthcare center for more than half an
hour past the time of appointment with the physician, and on other occasions
having been called at home a couple of hours in advance for the canceling of an
appointment due to doctors having stayed at home because of sudden sickness of
a child. And, opposite to patients, it is supposed and enforced that doctors,
as all public servants and authorities "obviously must" be
trusted, bypassing studies of trust in science and business, such as
Steven Shapin's A Social History of Truth, and
Lars Huemer's Trust in Business Relations.
The above account, based on personal experience,
should be completed with extensive reports of experiences like with the Swedish
healthcare application at 1177.se, matching
the national phone number 1177 for healthcare information services, as well as
with other examples that can be classified as HCI. The mentioned 1177.se is
said to protect the privacy of patients, who have to legitimate themselves by
means of a legitimation app furnished by banks and other entities. In
practice 1177.se formalizes and enforces which personal
health-care data can be inputted, made available, or communicated for which
purposes from and to patients, respectively healthcare personal in various
entities. Having missed appointments, pending variable rules, may be recorded
in the personal journal of the patient. Several examples could be adduced but I
refrain from doing so for the time being.
The healthcare example given above appears to be a
system which is being also used for surveying and correcting the
behavior or citizens. In this sense it can be related with one of the first
recorded historical investigations of the effects of computerization, Records, Computers and the Rights of Citizens (mentioned
below as RCRC, pdf-download here and here). In particular, the appendix F (pp.
247ff., but see also appendix E on Computerized Criminal Information
and Intelligence Systems). The title is Correctionetics:
A Blueprint for 1984 (written in 1973!). It has some comments stating
that systems for monitoring and control of the population of a prison
efficiently create a system with all the earmarks of the worst surveillance
data bank any civil libertarian could imagine. Referring to a study at that
time to "design a system to enable managers of correctional institutions
to make completely objective decisions about the treatment and disposition of
criminal offenders", it states that simple substitutions of the words
"governmental" for correctional and
"citizen" for offender in quoted excerpts from the
study transforms serious and humane objectives for prisoners into a nightmare
for citizens. The RCRC (p. 250f.) goes on noting that most of the
information relevant for computerized "correctionetics"
is entered in coded form. "The extent to which coding conventions match
the underlying structure of the data determines to a very great extent the
ultimate power of the computer program to handle any but the simplest sorting
tasks." It is worth noting that at the time many Swedes praised themselves
for being the first country in the world to have introduced a national personal
identification number (PIN), i.e. coding all
its citizens. The RCRC was a controversial American study of whether
such an innovation should be introduced in the USA, corresponding to what later
became the social security number. This matter can
properly be seen as political and cultural as illustrated in the equally
controversial Swedish study that I reviewed in a paper about Is the Swede a Human Being?
Superficially it may appear that the example above is
not representative for HCI as it is understood today, as manipulation of the
"interface" between a human and a computer. But the problem is much
deeper. The ultimate question is what is human, what is a computer, and
consequently an interface, relative to the task at hand. This will be developed
along the present text.
In the example above, coding corresponds to the class
"missed appointment", made equivalent to a "missed payment of
invoice" that, however, is framed in a historically grounded system of
enforced national (and international?) law. In terms of philosophy of science,
the problem of coding can only be the problem of "class assignment"
as part of the "teleology of measurement" presented by C.W.
Churchman in Prediction and Optimal Decision (chap. 5, pp.
107-110). It can be easily understood that no knowledge of these matters enters
into the design of the kind of computer systems we are referring to. In broader
philosophical terms the coarse statement on "the extent to which coding
conventions match the underlying structure of the data" portrays the question
of the relation of (syntax, semantics and pragmatics of) logic to
the real world. In this sense the whole matter of HCI reverts to the
discussions about logic and computerization as considered in my two earlier
essays on Computers as embodied mathematics and logic, and Computerization as logic acrobatics. The
limitations of logic cannot be patchworked by means of ad-hoc fragments of
HCI-models and gadgets, or by translating it into a question of "privacy"
that hides and avoids the core of politics and ethics as I tried to show in the
book (in Swedish) Systemutveckling och Rättssäkerhet [Systems
Development and Rule of Law], a simplified popular version of my doctoral
dissertation on Quality-control of information.
Or, better, the limitations of logic can appear to be patchworked
but the task is then to identify "invisible" capricious consequences
and to verify that they cannot be alleviated.
This issue has already been the object of considerable
attention, e.g. as early in 1972 at the time I happened to present the
dissertation on Quality-Control of Information,
and Hubert L. Dreyfus published his famous book What Computers Can't
Do, focused on "limits of artificial intelligence" - AI. It
was followed by an update in What Computers Still Can’t Do in
1992 (preceded by extended popularization e.g. in the journal Creative Computing, January
and March 1980) with shortcomings that could be inferred from the
later book's presentation by the MIT Press:
Today it is clear that "good old-fashioned AI," based on the
idea of using symbolic representations to produce general intelligence, is in
decline (although several believers still pursue its pot of gold), and the
focus of the Al community has shifted to more complex models of the mind.
The issue of HCI thereby appears to
extricate itself from Dreyfus' criticism, which was directed towards the first
wave of artificial intelligence - AI which as Wikipedia puts it was "founded as an academic
discipline in 1955, and in the years since has experienced several waves of
optimism". And we will see
also that it implied an as yet unmentioned HCI. It is as if the matter would
have shifted from "artificial intelligence" to "more complex
models of the mind", whatever that means beyond a continuous adaptation of
AI to the exploit of human interaction with more "complex" computing
technology. That is: what do complexity and mind mean beyond a repeated stated
pretension to understand and interact with or replace the human mind, a project
which is becoming to the fame of the Massachusetts Institute of Technology,
while encouraging criticism of technology that in turn is countered with derogatory terms such as neo-luddism and technophobia. It all turns into a squabble
with ad-hoc created neologisms, including "AI". All such approaches
tend to miss the point of getting to the core of the issue of what intelligence
is as related to humans and to reason, mind, brain, psyche or (symptomatically
controversial sort of synonym) soul. Dreyfus, for instance, relies heavily on
the philosophy of Husserl and Heidegger. I already have commented the latter in
several contexts that can be identified by going to the search-field (inserted in my homepage). The famous neuroscientist Antonio
Damasio tries to think deeper
and relies on Spinoza who is seen to have some important things about inquiry as indicated in Churchman’s The
design of inquiring systems (pp. 11, 25-27, 62, 69, 71-72, 261, 263, 282). Damasio borrows from
literature and psychology some very important, heavily loaded words like feelings and emotions, as I showed
elsewhere that also the
philosopher Kant does and abuses. Damasio does so without acknowledging the intellectual
struggles of the psychologist Carl
Jung, and symptomatically does not dare
to go into and still less define intelligence
(as related to inquiry). Those who
possibly expect Damasio to unravel the relation between brain, computer and
intelligence have not understood basic issues of philosophy of science, as
presented, for instance, by Abraham
Kaplan in his The conduct of inquiry (pp. 323-325 of 1964 ed.), despite of he himself ultimately being led
astray in the interpretation of Kenneth
Colby’s work as others are continuously
led astray to mirages by similar findings such as “Newly detailed nerve links between brain and other organs shape
thoughts, memories, and feelings” (Science, June 10, 2021).
And it is not only a question of mind
and such in relation to logic and computation. There is also que question of "interact" which subsumes the
meaning of reciprocal action. Without going into philosophy
proper, R. Ackoff & F.E. Emery's book On Purposeful Systems suggests (pp. 65ff.) that an action upon
an inanimate object (no human mind, no psyche, no soul) definitionally elicits
a "mechanical" reaction from an
inanimate object or an instinctually driven human, but a response from
animate subject(s) after a deliberation, choice or design of an
appropriate course of action or inaction which are expected to result in
appropriate ethically desired consequences. The idea is expanded in advanced
HCI by means of reference to particular schools of sociology, social psychology,
politics and culture, but there is no way to expect an integration of such
dimensions except in terms of a systems theory or overarching religion as
alternative to "ideologies". Or, as Wikipedia puts it, present day models, in general,
center around a steady input and discussion between clients, creators,
and specialists and push for specialized methodologies or
frameworks such as activity theory (whatever theory, models, methodologies and
frameworks are or should be). In comparison, the conception of Design of Inquiring Systems (chap. 3) proposes that a "steady input and
discussion" should be between carefully defined roles of (1) clients-customers-users,
(2) researchers-philosophers-designers-analysts-engineers, and
(3) decision makers-managers-executives-politicians. The latter are
being often neglected despite their representing economy-profitability,
politics and ethics in their relations to the clients, (4) shareholders-voters
who should but often do not include the affected rest (5) society-community-the
general public-people that transcends the narrow meaning of clients
and shareholders, and are today mentioned especially in matters of climate
change and global warming.
The decision-makers are those who
decide about the budgets and rules for use of the computer and, for instance in
the example above, about the rules for punishing patients like 80-years old
ladies with fines for the "crime" of not meeting doctors'
appointments, while doctors who stay away deserve to be trusted. (All this
while healthy researchers and philosophers deal with the issues in this present
text). When she "interacted" with the computer at 1177.se in order to
book an encounter with the doctor, she could not know that an ignored somebody
else like a clerk, on the basis of rules changed by a third deciding party,
would later also "interact" with the computer in order to use her
data for charging an unexpected fine that ignores the circumstances of it all.
In other words, all the possibly involved people should know which are the
changing "circumstances" of the possible use of the computer and its
consequences for themselves and others. "Knowing" this can be
conceptualized as everybody, but especially the clients or stated people
serviced by the computer and its software, should have and be able to keep in
their minds a "user model". All this is at the same time trivial and
theoretically-practically impossible, except for (just, paradoxically) trivial
cases, and especially when economic-political factors are absent, as some
people trying covertly under the cover of complex software and interfaces to
gain power or profit at the expenses of others while appealing to
"justice". One main point, however, is to realize that these problems
may be ignored in HCI and be attributed to and the responsibility swiftly
shifted to the field of social informatics and "sociotechnical interaction" where to my knowledge Rob Kling was
one of the few who cared for similar problems. This shift, however, undermines
the scientific legitimacy of HCI since the facile reference to "user
models" mentioned above disguises and avoids the basic problem of the
limitations, dangers, and consequences of computation, computers and therefore
of interactions themselves.
A way to explain the theoretical difficulties of
interaction at the level of philosophy of science is to realize that the idea
of interaction implies often if not always that a unique somebody interacts,
implying "uses", and uses "something" such as a
"tool". It is important to notice that the same idea appears under
the label of computer “support”. I tried but failed to capture this idea in a
project about Essence of computers: presuppositions for support returning
to it in Expert-support
systems: the new technology and the old knowledge.
Its only justification is that by considering the computer to be a tool, the
hope or illusion is conveyed that the human may intervene in correcting or
completing the shortcomings or errors of the computer, if one knew what “error”
is, and consequently also knew the difficulties behind the conception of human error
and human factor. The core idea of this expression, tool or support, it
should be emphasized, is that the tool is under the full control of an
all-knowing user guided by an
all-knowing manual which is also a tool and with a necessary experience
or tacit, implicit knowledge, which is not
equivalent to “unreliable feelings”. Never mind tool-makers or engineers
beyond the user have been doing or meant. And it happens to be also
an idea that is or should be subsumed in the expression "human-computer
interaction".
This is viable so long as people have not understood the controversial
presumption that a computer is so "intelligent", whatever
intelligence means, that it can be equated to a human. Then it should feel
controversial to "use" another person as a tool, even for those who
do not care of Kantian ethics of practical reason and the categorical imperative “Act so that the maxim
of thy will can always at the same time hold good as a principle of universal
legislation”, implying the prohibition of treating humans merely as a means. It is, therefore, felicitous if computers neither
can nor should be conceptualized as human minds, whatever the often
misunderstood “mind” means beyond misunderstood trivial logic. Nevertheless it
may be already too late: infantile computer nerds and marketers are already
building upon the idea of autonomous AI (whatever
programmed autonomy is) including military robots, micro-drones,
and the rest as if it were only a question of supporting campaigns and research
for stopping killer robots, or academically: “to ban lethal autonomous weapons”. This despite of the entertaining presentations of dancing
and singing humanoid robots, not always within
their explicit context of "Army
Technology". The whole is
then embedded in academia as represented, for instance, by professor Stuart
Russell with a puzzling new perspective of HCI as at the Center for Human-Compatible Artificial Intelligence at the University of California. Note: Human compatibility, a new quagmire to be compared with e.g. Human-centered artificial intelligence elsewhere. Humanity (or
was it humanism?) as in HCI, seems to be much in vogue, albeit its misuses
already had called my attention in a paper so early as 30 years ago on Computers in human science or humanistic computing?.
It is therefore quite unfortunate
when HCI researchers as commented in Wikipedia
observe the ways in which humans interact with computers and design
technologies that let humans interact with computers in novel ways. The term
connotes that, unlike other tools with only limited uses (such
as a wooden mallet, useful for hitting things, but not much else), a computer
has many uses and this takes place as an open-ended dialog between
the user and the computer. The notion of dialog likens human–computer
interaction to human-to-human interaction, an analogy which is crucial to
theoretical considerations in the field.
It is unfortunate, because it
is not such an analogy that is crucial to “theoretical
considerations” in the field, whatever is meant by “theory” and “field”. The
computer “has many uses”, but it is problematic to understand what “use” and
“user” is, and even more problematic to know what “legitimate and good use” is.
Therefore it is also problematic what the “grand old(est) man” Terry Winograd means in the book co-edited with Paul
Adler Usability: Turning Technologies into Tools (1992), contributing to perpetuate the idea of
the computer as a used “tool”. The
notion of dialog should not liken human-computer interaction
to human-to-human interaction. It is entertaining in this context to note what
V.A. Howard and J.H. Barton write, quoting D.N. Perkins, in their
remarkable book Thinking on Paper (1988, p. 69) that can
be seen as dealing with “dialog” between writers and readers (not tools)
through HCI, indicating that HCI, on the contrary, can “freeze” knowledge with
its premises and suffocate true reasoning:
Formal principles of logic are of course implicit in everyday reasoning,
but like grammatical rules, it isn’t necessary to know them to use them. […]
“Effective reasoning…depends on an active effort to interrogate one’s knowledge
base in order to construct arguments pro and con” […]. “In contrast with
reasoning as classically conceived, premises change and accumulate as the
argument proceeds rather than being given at the outset. Second, the premises
are somewhat constructive, elaborating the reasoner’s understanding of the
situation rather than merely objecting”.
The problem, however, lies already at
the root of it all in that the computer is not basically a tool, but in the
best case it is an instrument if the instrument, as in the
case of measurement, is not for mechanical direct manipulation but is
theory-laden. It is the case of scientific, e.g. optical, or even advanced
musical instruments. This stands as the basis for understanding the possible
meaning of what is said about measurement in quantum physics, and of an otherwise controversial quotation from
some famous remarkable "rejected parts" of the dissertation by the mathematician Jan Brouwer ,
also reproduced in an earlier essay of
mine on computers seen as embodied logic and mathematics:
[T]he
physicist concerns himself with the projection of the phenomena on his
measuring instruments, all constructed by a similar process from rather
similar solid bodies. It
is therefore not surprising that the phenomena are forced to record in this
similar medium either similar laws or no laws. For example the
laws of astronomy are no more than the laws of our measuring instruments when
used to follow the course of heavenly bodies.
This is the reason why an ambitious international
symposium in Sweden in 1979 resulted in an inconsequential hodgepodge of
disparate opinions and insights, reported the following year in Is the Computer a Tool?. Nowhere
there is a focus on the essence or nature of logic and mathematics and their
consequences. Even an insightful participant of the symposium such as Sherry Turkle succeeds
only in rising (p. 88) the question of "what computers
really are", but regrets the missed "epistemological
issue: computation is irreducible". The epistemological background and its
limitations can be inferred from a paper she wrote in 1992 with her husband and
("symptomatically") AI-oriented computer scientist Seymour Papert, "Epistemological Pluralism and Revaluation of the
Concrete" in Journal
of Mathematical Behavior,vol. 11 (1). But:
and so what? Computation reducible to what,
what could it be? Computation of what and what for? Numerical and, or, logical
computation? Is computation (what kind of) mathematics or logic? Which are the
relations between computation and…what else? And, therefore, which is the value
of computation, by whom and for whom? Who cares for the question of whether the
computer is a tool or not? Will we see a Kantian reduction of this issue to
aesthetics and art?
The following quotations (pp. 16-17)
from the paper mentioned above may give a taste of its analysis adapted to
feminism or, rather, to "women studies" on the basis of psychoanalytic theory, as contrasted to analytical psychology in my old introduction to the study of Reason and Gender:
When women neutralize the computer as "just a tool," it is more than
a way of withdrawing because of a lack of authenticity in style of approach.
Insisting that the computer is just a tool is one way to declare that what is
most important about being a person (and a woman) is incompatible with close
relationships to technology. [...]
Achieving epistemological pluralism is no small thing. It requires calling into
question, not simply computational practices, but dominant models of
intellectual development and the rarely challenged assumption that rules and
logic are the highest form of reason.
The most important latter question is then what the "computational
practices" or the computer is, if not "just a tool", or a tool
for discussion about differences between women and men. Turkle's (and Papert's)
approach in their paper "solves" the problem in a "Kantian
spirit" as I suggest in my own study of Computerization
as logic acrobatics.That
is, it all ends in talk about art and artful integration between genders, seen
as an integrative reason where reason remains being plain logic, supposedly
analyzing the human unconscious, all consistent with the limitations of
psychoanalysis itself. This seems to be confirmed by the summary of discussions
written by Bo Sundin, member of the symposium committee, in his editing of the
report Is the Computer a Tool? (p. 128): researchers should not
only be scientists, but by having scientists and artists present at the same
time one could overcome the limitations of science; the academic community
could profit from contacts with artistic institutions, and plays, literature,
etc. could be used in the discussions. I myself consider, however, that his
most important summarizing observation was instead the following (p. 127):
The complexity of
modern society is to a large extent due to the lack of traditions like
religion, which in older times acted as an organizing principle. The attempt to
install computers to deal with the unorganized, incoherent complexity of our
society is harmful in two ways. It increases the complexity and avoids actual
problems.
And it is not easy to see why a Kantian reduction of the question to
aesthetics, art or to feministic epistemological pluralism should be of any
help for the design of interaction with such tools or instruments, except for
making them more marketable, profitable or funny to deal with. It will not do
to try to counter the references given above by programmatically declare them
obsolete if they are older than 10 or 30 years, set against the extraordinary
rapid development of computer and communication technology. Without cumulative
knowledge or building upon old knowledge, even if for no other purpose than to
show what was false, there would be neither science nor technological
development (not to mention the problem of generation
gap). This is also the message of C.W.
Churchman's last book on Thought and Wisdom (chap.
3 on "Success of failure"), which happens to
illustrate the core idea of my own dissertation on Quality-Control
of Information. And now, let's see
further examples of human-computer interaction, mostly of the apparently most
trivial sort that almost escape the common HCI-technicalities until a last most
complex example that exceeds such most common technicalities.
Example
#2 - Telecommunications company
One main Swedish
multinational telecommunications company and mobile
network operator illustrates modern trends in the provision of customer
service. A customer gets on her mobile phone an automated warning that her
subscription to the services ends the following day. She wishes through her
phone to make a "semi-automatic" renewal of her periodic monthly
services-subscription for the sum (99 SEK, Swedish crowns) that is inferior to
her remaining savings at her account at the company totaling a rest of (111).
She dials an SMS to a given special number of the operator with the coded text
"Fastpris 99" [Fixed price 99] but she gets
a computerized answer that it is not possible because the required amount has
no coverage in her account. That is, does the system calculate that 99 is more
than 111? Her choice is then to dial and check again that her available amount
is 111 and repeats the request for the renewal that it is again responded
stating that she has not enough money left. The operator's site does not allow
for digital or postal messages to people and it is evident that it does not
want time-consuming dialogs or questions requiring human answers. The exception
is for at a particular phone number of the operator where there are usually
waiting queues of up to 30-40 minutes. The lady tries this and her choices
among 3-4 steps of alternative codes are successful in that luckily this time
within 5 minutes she gets into contact with a customer serviceman who, however
does not know what to do and says that he cannot forward this customer's
complaint neither to software experts nor to management. He advises her to try
again the same procedure on the following day (when her subscription is no more
valid) or to try other paths at the company's internet site. The same day the
lady tries again starting again at such "other path" at a site
arriving to an offer of choice among a series of pre-formatted customer
problems. Choosing the closest one to the actual problem the lady arrives to an
offer for "chat" where she previously for other problems had been
able to chat with a serviceman. After typing a description of the problem she gets a strange chat-answer offering to
click and choose among several pre-formatted problems that have nothing to do
with his issue. She gives up for this first day. The problem up to this point
obliged the lady to call once the operator company's only available phone
number for direct human voice-contact, followed by an inquiry into an internet
site with hope for a human chat. The call required several steps of choosing
among different digital codes for different services, fulfilling obligatory
electronic legitimation, followed by phone queue waiting time of about 10
minutes until human contact with a customer service employee of the operator.
The Internet inquiry required the same except for a personal legitimation.
The following second day, after her repeated
explanatory reports of what it was all about, that she was not allowed to use
her credit amount in order to renew her monthly subscription, the second
customer serviceman, in the guise of goodwill compensation, inputted further
(the company's) 20 SEK to her credit than then amounted to 131 SEK and asked
her to repeat the loading operation for 99 SEK. Nevertheless, repeated trial
also failed: are 99 SEK more than 131? She calls customer service a third time
and explains herself - always to a new different serviceman at the company's
call center. He consults an expert technician, promising that he or somebody
else would call her back. It does not happen (prolonging the story to the point
of exhausting the readers and the writer of this present text…) The lady,
repeating the same procedure the following day for the fourth time, was finally
informed that the personnel of the customer service had been just informed that
the operator company had experienced bugs in their systems. The man at the
customer service offers to supplement the account at the expense of the company
with another addition of 99 SEK corresponding to a one-month subscription
free-of-change. It is done, but the lady's account totaling finally 230 SEK
could still not allow a prolonged subscription. Repeating the procedure once
again, for the fifth time, now after waiting 10 minutes a phone queue of 33
people, and arriving at being serviced, an automatic voice informed that the
final connection to a human had failed and she had to start again, waiting this
time also 10 minutes in a queue of 27 people. After a repeated explanation with
a fourth serviceman who investigated the case for about other 10 minutes, she
was at last informed about the mystery. When her account balance was 111 and
she wanted to use 99 for a renewal of the subscription it was not allowed by
the "system" because 20 out of the 111 consisted of an unperceived
complimentary bonus she did not know about, given to her by the operator
company because of something she didn't know – but that according to an unknown
rule of the company could be only used for expenses outside the
conditions-limits of the subscription. For instance, as a charge for surpassing
the allowed monthly 1 gigabyte of allowed surfing. In other words: the account
balance consisted of two unspecified balances to be used for different
purposes. And the lady reflected that her feelings where akin to have been
robbed and "raped" in the sense of abused by being obliged to loan
her body and mind for a total of at least four hours in order to request to be
allowed to spend her own money, being put in a sort of straitjacket. The
solution from the beginning would have been for the lady to deposit at least
the difference between 111-20 = 91, and 99, that is to deposit at least 8 SEK
or to never mind about why and how (as many clients are tempted to do) by
deposing on her account a massive sum of, say, 300 SEK, hoping for the best. It
had not helped that the customer service officials had tried to deposit
complimentary additional amounts of 20+99 up to 230 since they both were
automatically and unknowingly classified as "bonus".
The same company also offers another example when it
wishes to reimburse the customer for some amount that exceeds the estimated
costs of some service. The customer receives a communication that a certain
amount will be reimbursed, but if he happens to follow up and discover that the
reimbursement is not made and asks for an explanation, he is informed that he
must return to the company’s site in order to, after waiting in a phone-queue
for up to 30-40 minutes, identify himself with a special app, and finally
communicate the number of his bank account to which the reimbursement can be
deposited, despite of this number having been already declared known by the
company because it had been used for automatic regular bank payments of
periodical fees. When this does not have any effect, an investigation of the
personal digital track of invoices at the company’s site reveals that the
announced reimbursements had already been automatically used by the company for
advance payment of new wrong invoices that must in turn be questioned. And so
on.
Reflections upon example #2
I mentioned above that the lady has to start the
process all over again, repeating her text and getting strange answers until a
close relative trying to help her understands that she has been dialoging with
a chatbot. The
relative was an older friend, old enough for knowing what a chatbot is, and
knowing its historical precedents in computer science, associated with computer
scientist Joseph
Weizenbaum and the famous ELIZA would-be-AI software. Finally, reflecting
about above mentioned "obsolete" 10-30 years old knowledge: today (September
2020) the Wikipedia declares that ELIZA was "Created to demonstrate the superficiality of
communication between humans and machines", but I remember that it was
instead seen as a step in a glorious development of AI. Today it is used as a
perfectioned reality available to customers, thanks to more advanced computer
and communication technology. In August 2022 the world press started to spread
news like in The Guardian (9 August 2022)
about the Facebook-META’s “BlenderBot”,
a chatbot
that is supposed to allow anybody’s mind-blowing conversation with AI-computers
about “anything”. Its dangers may direct our attention of the type of
literature on “sick society” after R.D. Laing’s
work. I myself prefer Carl Jung’s reflections on Civilization in Transition (Collected Works CW10). Weizenbaum himself had gone meditating about the for him
surprising positive response to ELIZA's fake conversations, and finally
wrote an exciting but inconsequential critical book Computer Power and
Human Reason, in the same principled and righteous but powerless
intellectual style as Turkle and Papert above. Weizenbaum launches a gratuitous appeal to the "difference between deciding and choosing"
and deplores computers' "lack of human qualities such as compassion and wisdom",
while Wikipedia in this context ambitiously refers to this contribution's
relation to the "ethics of artificial intelligence"
(and human-computer interaction?). For the rest, it turns out that the
"logic" of HCI of the system presupposes that the clients somehow
know from somewhere that their account balance consists of two unknown parts
and this balance can be used for different purposes according to temporary
unknown aspects of the pricing of the operator company. The rest of the example
does not deserve further comments.
Example
#3 - Telephone answering machines
We have seen in example #2 that one main feature of
modern human-computer interaction, a feature that is inherent to the essence of
computation and paradoxically even to (computerized) communication seems to be
to try to minimize (costly) contact of customers, clients, or citizens with
(for management, owners or supposedly taxpayers) costly employees in business
or government. Once upon a time when somebody called a phone number to an
absent employee there could exist a secretary who
would annotate the caller and transmit the message. Then came telephone answering machines where the
caller left a message, often proposing to be called back at a certain time.
Phone answering machines could be completed with built-in messages telling at
what time the person could be found, or else that would divert the call
relieving the called person from any further responsibility. The main
"improvement" from the called person's perspective appeared to be to
relieved from further responsibility like having to call back all those by whom
he had been called, and possibly to discourage people to call at all. The next
step became to enumerate most if all the possible question that could be put by
the caller and to automatically, by voice recognition or choice alternative
dialing codes, redirect him to a series of pre-recorded answers, other numbers,
or internet addresses. The "final solution" seems today to be to
discourage or outright deny the possibility of calling at all. For business
this implies the risk of losing potential customers and to certain extent can
be obviated by denying the possibility of calling sales personnel but allow the
call to be directed to a central human secretary or call-center with competence
for answering questions. This is, as we saw, at the risk of the caller
incurring into the above-mentioned event. For government and public agencies,
however, one final possibility is to outright deny the possibility to call any
human agent. The ultimate idea is to save what in the affluent West is
considered most costly for organizations: manpower, yes for organizations, forgetting the counterpart of unemployment
as I show it is forgotten
in the context of implementations of artificial
intelligence, as
it is also forgotten in the context of stress for paradoxical lack of time for
increased productivity, as analyzed by Staffan Burenstam Linder in his Den Rastlösa Välfärdsmänniskan (originally written in English, as The Restless Welfare Person).
And this is often done by means of computerization. It
deserves a special
analysis I have already attempted to do,
and with minimal consideration for long-term social and political consequences
starting with focus of the further consequences of the “profitability” of
savings of manpower, as studied by, say, (in Swedish) Christer Sanne's Arbetets Tid [The Time of Work] and Keynes Barnbarn [Keynes'
Grandchildren], and more recently Roland
Paulsen's Empty Labor. It reminds the
promise of shortening the waiting time of customers-clients at the checkout
queues, starting in supermarket chain stores, thanks to the introduction
of International European Article Number (EAN),
this without considering that the result would be not the shortening of
queue-times but the reduction of the number of cashiers and the transfer of
cashier tasks to self-serving customers at automated cashier terminals. It is
an example of what has been called as the trend towards Heteromation, or
a new “division of labor between humans and
machines, shifting many people into work that is hidden, poorly compensated, or
accepted as part of being a “user” of digital technology”.
The additional example promised above is related to
the previously mentioned site 1177.se. It
is easy to understand that physicians, at least those who are employed and paid
in the Swedish welfare system, do not wish to get phone calls and e-mails from
patients (who are not really "customers") since this would increase
their workload. I myself did not know how taboo such a practice was, since some
physicians did approve such simple communication, e.g. for being informed about
results of ongoing treatment. I was shocked, however, when a heart specialist
who was going to treat me for an emergency crisis of high blood pressure
suddenly disrupted our relation and canceled a coming appointment, refusing to
treat me further – just because I had sent him a short e-mail in advance in
order to prepare for the later consultation. This event motivated my notifying
the authority for monitoring of medical ethics, since a delay in treatment
could put my life at risk in case of stroke. The history of this notification
is documented in a summary available on
the net. The case of 1177.se, however
illustrates the trend in the perspective of HCI since it shows the increased
burden of the patients sometimes seen as clients or customers - in their
relation with doctors/physicians seen as decision makers, managers, public
officials. Contacts are generally obligatorily channeled through 1177.se under
the assumption that it guarantees privacy required in medical contexts. It
requires obviously a computer or cell-phone, a login supported by an electronic
identification-legitimation (which in turn has required logins and
legitimation) all based on PIN personal identification number (commented
above), followed by a series of HCI-steps consisting of choice of coded text
that identify the relevant health care center, hospital or clinic. Arriving at
the latter there are generally neither email-addresses nor phone numbers, and
there may be or not be a choice about e.g. request for a new or cancellation of
an appointment, renewal of drug prescriptions, reading a selected texts out of
one own's medical journal, or of writing a message of a maximum of 2000
characters to a certain person or kind of person whose answer will be announced
later per an SMS to have become available, but with a text that can only be
read at the 1177.se site after a renewed login. My highlight, however, is to
remark that certain clinics do not allow to send messages. Such
messages may require qualified manpower such as doctors who are expensive or
not available.
Reflections upon example #3
Beyond some preliminary reflections already contained
above, let's recall that once upon a time there was a street or postal address
(that today and are seldom advertised) to which one could write a letter with
any number of characters at the sender's risk, could put in an addressed
envelope with stamps, and mail it. Nobody could stop one to do this, and at the
letter's arrival somebody had to take the responsibility to throw it away
unread. But recently when I had no other way to question an invoice of a telecommunications-operator,
I had to send a registered conventional mail to the given postal address at a
cost of about 10 US$, it was returned to me after a couple of weeks because the
operator-addressee had moved to another postal address without a "forward".
The invoice was on paper but if, as suggested, I had chosen digital invoices to
the Internet-bank I would not have had any available postal address at all. The
main reflection on 1177.se, however, beyond those already implied above is that
HCI purports to facilitate the "use" of the computer but what happens
is that it forces the human thought and behavior to adapt to the
(logical-technical) requirements of the computer or the requirements of those
who decide the introduction of the computer for tasks that earlier where performed in other ways. The basic assumption of
computerization and its HCI is “profitability”, to save money of/for specific
entities, mostly by saving specific kinds of manpower, either to do more and
faster with the same amount of manpower, or to do the same with less manpower.
The example only illustrates one case in which the customer, so to say,
"HCI-worked" up to 4 hours for "nothing": a case of
technologically conditioned "empty labor". And when this is said, we
leave aside the question of who's profitability, labor vs. capital etc., since
it would require, if not Marxism at least complex PPB (planning, programming,
budgeting) evaluations whose complexity already is obfuscated at the old level
of the seventies' in Roland
McKean's Efficiency in Government through Systems Analysis that
before of becoming New
Public Management and being applied,
among others, to Swedish universities was deadly criticized as
early as 1983 in Ida Hoos' Systems Analysis in Public Policy.
Example
#4 - Personal identification number PIN
At least in Sweden – on the basis of the early and
extensive use of PIN, personal identification number, there have been cases in
which authorities have sent away wrong communications to relatives and others
about the death of a certain person, sometimes to the still living person
itself. I have neither kept through the years cases reported in the press nor
been able to retrieve documentation on such events except of one example (in Swedish, in Aftonbladet 26
Nov. 2009 & 11 March 2011) where an adoptive son received such a
communication about a father who had died several years earlier, entailing
serious repercussion. There are analog examples of authorities' registers
classifying a certain person having a (in Swedish) "betalningsanmärkning" [payment
remark, owed money] which may be questioned, a
(Swedish) example being found on the net.
Reflections on example #4
In Sweden the sheer existence of The Swedish Data Protection Authority allows
a certain documentation and understanding of the complexity in correcting some
data existent in the great number of databases. What is worth of reflection is
the difficulty and the amount of labor (including and exceeding HCI) required
for correcting wrong information that in belonging to a database can in a short
time lead to multiple consequences, which in turn have to be corrected manually where
"manuality" is a brainy question and is
hard to define. The difficulties are comparable to the earlier example (above)
about the client of an Internet operator with the difference that they may have
serious juridical and economic dimensions.
Example
#5 - Public transports and parking
About public transports and parking
tickets. Once upon a time the passenger paid for a ticket and got the ticket in
his hand, as a concrete visible proof of the paid amount. Computerization using
magnetic tickets or cards that are loaded with certain sums of money imply that
the control of what is loaded on the card and what is spent on each trip
occasion is left to the "system" about which the user or client has
neither understanding nor overview-control. Information about the transaction
appears on some screen. It appears only for a fraction of a second that often
is difficult to grasp, or it can ex-post be identified by a reader available at
a ticket office. The transportation company itself has ruled that the card
itself may lose validity after a certain time, so that in order to use the
remained credit amount the customer who did not memorize the expiration date
may have to step down and walk to a certain office for renewal. It happened to
me. Or the magnetization is lost by contact with a strong magnet or cell phone
as the case is with magnetic keys-cards at hotels, and so on. Furthermore: on
occasion of the covid-19 pandemics the public’s already paid digital credit
cards for loose tickets (“reskassa”) were declared invalid until further notice, after October 28th 2020 in
order to offset the company’s accumulating budget deficit. Declared
invalid, despite of the traveler having already paid cash, in
advance, a whole amount to the transport company corresponding to trips
that he may not yet have done, that is, the paid tickets not yet having been
“consumed” by the traveler (as in my family’s case), while the company can
already have consumed the traveler’s money. Everybody including old
retired travelers were obliged now in covid-times to buy the ticket
somewhere in advance of every trip, or if no “somewhere” was available at the
start of the trip, they had to have a mobile feature phone (which
many elders do not have) with which to pay electronically their single-trip
ticket (an operation that many elders have never done) after electronically
identifying themselves (operation many elders never have done), all pending a
fine corresponding to about 150 dollars, after an eventual ticket checker’s own
judgment. All this motivated by the economic losses incurred by the public
urban transport company because of the unavailability of digital readers other
than at the usual front entrance of the buses that were barred in order to
protect bus drivers from covid-contagion, and after that the task of
former ticket sellers was taken over by drivers, and later further taken over
by digital readers of digital tickets under the supervision of the driver,
readers that are unpractical to move. Apparently, an accumulated final burden
of failed man-machine interaction is forced upon the client-population in
general and elderly in particular who for the rest are advised to avoid
traveling in the dangerously overcrowded buses.
This transportation case is analogous
to payment of parking charges. From the beginning one could put coins in a
parking automat, and later one could pay with coins or credit cards in an
automat, getting a receipt for the paid amount and time, plus a paper ticket to
put under the car's windshield, stating the valid paid time. Ultimately, the
client gets no validation except for a statement on the screen of a ticket
machine where his credit card pays a certain amount for a stated time that he
has to memorize, on the basis of a license plate number that, by the way, can
get some character wrongly digited leading to a fine. The process is further
complicated by the introduction of parking-apps or mobile phone applications
developed by a plurality of parking companies who share parking lots in one
same region such as a city. And they require from customers, car owners,
further laborious updates of their parking apps for enabling their continued
use based on periodic debugging or innovations. All this
still disregarding cases such one I know of an elder in cold and
windy weather who upon arrival was refused to pay with their credit card in an parking automat because it was out of order. And he got
confused by having to interact with two different screens of the automat at
consecutives stages of the interaction, one for choice of estimated parking
time and the other for validation of his credit card. An app for alternative
interaction in place of the automat, which was specified in the displayed
instructions, could not be downloaded to the cellular because of occasional
poor coverage. He tried in vain to call the displayed phone call number for the
responsible service company for communicating this and preventing being fined,
but after choosing among a series of coded alternatives was requested to wait
his turn in a queue, for an undetermined period of time, the whole totaling
more than 15-20 minutes, while he was in urgent need for a rest room.
Reflections on example #5
Some reflections on the urban
transportation were already included above because of stylistic editing
reasons. Concerning parking, the client has no proof in his hand and often has
to memorize the valid parking time. This is so if the parking ticket machine
(now with virtual tickets) works properly. If it is noticeable that it does not
work properly, the client is requested to call at his own expense a certain
phone number for reporting the dysfunction. If it is not noticeable, he can
only hope and “trust” (whom?), but reclamations ex-post, after getting a fine,
are hopeless without any proof on his hand. It recalls the case above in this
text, of trying to deserve consideration and justice vis-à-vis an internet
operator. The main reflection here is exactly this loss
of control and power by the client or customer who is punished, fined
and made responsible for further corrective action, posited that this
client-customer identification is possible. That is, in general different
classes of people may be considered as clients or customers, not to mention the
synonym "users", including engineers ("researchers"?) who
get their salary for designing the computer and communications systems. All
this similar to the fact that in the political process there may be different
(hierarchical?) classes of decision-makers, politicians or CEOs inside the
computer and communication industry as well as inside the military-political
establishment. One can also say that power is handed over to the
"system", so far as one is not obliged to specify how a system is
defined, if not in terms of a Design of Inquiring Systems, which in
turn risks to "pass the buck" to a "combination" of other
disciplines. But it is also possible to see that power is being surreptitiously
passed to an anonymous system of "cognitive logic and mathematics"
formally in the hands of a certain class of people, akin to what Edmund Husserl criticizes in his, symptomatically
unfinished The Crisis of the European Sciences. Especially
in its second part about the depletion of mathematical natural science (cf.
computer science) into its "mechanization", but leading to a dead end
of phenomenology because of a failed critique of Kantian psychology. More on
this in my earlier Computerization and logic acrobatics.
Example
#6 - Banks safe deposit box
Once upon a time if one wished to get access to one's
bank's safe deposit box it was enough to go during the opening hours to the
bank's branch office with an identification document. Some year ago the bank mailed a letter to its customers stating
that starting a certain date it would be mandatory to call the branch one
working day in advance to announce the wish to come the following day. No
explanation or reason was given. On occasion I realized that it was no longer
easy to find the branch phone number. No yearly paper catalogs are anymore
printed and distributed, only sites and data bases on the net with numbers that
direct the calls to call centers in order to save the time of employees who
would have to answer phone calls at the branches. In some other contexts call
centers are even outsourced to cheaper foreign countries where those who answer
have even more limited detailed knowledge of the local language and
environment. But in my case I called the only available number of the
bank's headquarters call center where I was put on a waiting queue of 55 people
with a waiting time of 25 minutes in order to get a phone number of the branch
that I called and after a digital identification on a special app I could
register my wish for a visit the following day in order to get at my safe
deposit box.
Reflections upon example #6
Once again, this is an apparently trivial case that
can be seen as a digital transfer of labor to the client, possibly motivated by
increased risk of robbery that lately has increased with arms and explosives,
to be expected in the context of safe deposit boxes. It is also the case that
such developments are attempted to be provisorily countered
with digital controls that affect customers.
Example
#7 - Dental hygienists and antibiotics
A man was called for a periodic treatment to a dental
hygienist. Having incurred earlier in a sepsis caused by dental infection and a
consequent heart infarction operated at a hospital, he had also been
recommended if not outright commended by the hospital's medical dentist to
prepare for future dental hygiene with a complete antibiotic prophylaxis. When
he explained this to the dental hygienist, the responsible dentist refused such
prophylaxis on the basis of guidelines from the public health authority intended
to combat societal antibiotic
resistance. Other medical authorities at the
hospital observed that antibiotic resistance implies a statistical point of
view that must be revoked on the basis of individual considerations. A human
life cannot be sacrificed on the basis of such statistical prevention measures.
The local dental hygienist, however, had no access to the hospital's digital
patient journal, and an inquiry directed to the hospital's medical dentist was
responded by a new managing consultant doctor by directing the dental hygienist
to those at other departments of the hospital who had taken care of the
historical sepsis. The local dental hygienist, however, had no technical access
to the hospital's digital patient journal and no prescription for antibiotic
prophylaxis was allowed. The risk was taken by the elderly patient who up to
date has survived.
Reflections on example #7
This one additional example of reliance upon digital
computerization and communication that breaks down at the interstice between
different organizations, shifting the burden of responsibility, communication
and risks of broken communication upon "others", and ultimately to
the client or patient.
Example
#8 - Radio Sweden’s summer program
On August 11, 2020, the public service Sverige Radio [Radio
Sweden] broadcasted one of the daily summer radio programs
in a series “Sommar i P1” [Summer in
Radio’s Channel 1] where selected, often well known, personalities are allowed
to talk, alone by themselves for ninety minutes, about their lives, opinions
and interests. In this program a young woman, a hockey player, used a part of
the program complaining of harassment by a man who was the head coach for Sweden women's national team. A listener perceived
that the program contained a personal attack to a person and his organization
despite the official requirement of Radio Sweden’s impartiality and its advertised rules, and wished to direct a complaint to
the Swedish Press and Broadcasting Authority (here
abbreviated SPBA). It directs the public to its Internet site where there
is an offer for complaints through its e-service requiring an initial digital
authentication of the person’s identity, which is followed by a detailed guide
with requests of details about the plaintiff and the program in question, plus
his motivation of the complaint. This was done in about half an hour but when
the whole digital form was meant to be sent away nothing happened. After
repeating the process with all the fill-ins nothing happened again. The
plaintiff then copied all the text on the computerized form and keyed it in a
digital text document that he sent by e-mail to the official main e-address of
the SPBA, together with a complaint that the digital form input did not
work, after having spent more than 90 minutes for the process. After about five
weeks he received per regular mail a communication (ref. 20/03658) that “The
Review Board has reviewed the current program in the light of the notification.
According to the board's assessment, the program does not contravene the
requirement of impartiality.”
Reflections on example #8.
The whole effort resulted in a three lines’ answer without any
motivation or comment. In this respect the response from SPBA recalls
what I already have written about the “repressive
tolerance” in the context of complaints regarding
medical safety of patients. It is the arrogance of power, now dressed in the
digital HCI-form. The public is given the opportunity to believe that it can
complain, as an escape valve. It hides a mechanism of requesting an ultimately
exhausting investment of energy leading to a mute discouraging answer from a
formal “authority” that in our case was represented by the signature of the
decision document by a member of the Supreme
court of Sweden (appointed by the government and
member of the feminist women’s network HILDA)
who was given a part-time assignment in judging SPBA matters. This is
hoped to be enough for the public’s acceptance of this type of answers from
authorities working for other authorities. For the rest the requirement of
impartiality means, according to the Review Board's practice,
that if serious criticism is directed at a clearly designated party, the critic
must be allowed to respond to or comment on the criticism. As a rule, this must
take place in the same program or feature, either by the critic himself or
herself or by reporting a comment from him or her. The event was the more
remarkable because in most if not all other contexts the radio or tv program
producer, host, anchor, or moderator very promptly interrupts and silences
whoever is speaking with the above mentioned motivation, the more so in
questions that may be subsumed under the Swedish
law against discrimination (SFS 2008:567, cf. Act concerning the Equality Ombudsman) or
questioning the democratic order or the radio and television act (pdf, SFS 2010:696). At the same
time various documents remind that constitutional right to and limits to freedom of expression, related to the constitutional fundamental law of freedom of expression (pdf),
one of the four fundamental laws of the Swedish constitution.
The inherent complexities become more visible in both national and international media’s repercussion of the case of
the Swedish professor of geriatrics Yngve Gustafson who
also was invited to have his summer radio program in the same series in Radio Sweden
on August 13th 2020. He is referred, in contrast to our reported case
with the hockey player, to have had his text censored by Radio Sweden in
what regards his criticism of health authorities’ treatment of elders
in the covid-19 pandemics. Quotation from reddit.com: “But
for me it became inevitable to talk about the corona storm as well, he says.
But I was not allowed to talk about the authorities' ignorant and unethical
actions and I do not think that feels good, he says”. Statistics with
all its uncertainties show that in October 14th 2020 Sweden was the fourteenth
among 216 listed countries of the world in terms of number of covid-19 deaths
per million inhabitants, more than 583 (mostly elders), well above other
northern or Scandinavian countries Norway (51), Finland (62), Denmark (116),
and Iceland (29).
Let us emphasize why this all is worth of consideration in the ongoing
reflections: the seriousness and importance of freedom of expression is what
motivates the strong (HCI) requirements in its mentioned questioning. At the
same time this is not at all respected in the authorities’ answering or
countering of such questioning. In terms of HCI “action-interaction theory”
there is only digital action but no substantial reaction or
response, and this is the nature of power over dialog. The
more so if the case happens to be about a feminist hockey player with
LGBT-connotations who in a spirit analog to #MeToo accuses a male coach for gender
discriminating harassment, in a country where government has engaged itself
in Sweden having the first feminist government in the world. It
is easy to imagine the design of a digital form for response to a complaint
which requires, for instance, at least the filling of a blank field for
“motivation” for rejecting the complaint, a motivation beyond that the original
complaint was just unmotivated. If this is so in such naively simple technical
contexts it is easy to imagine what happens in more complex ones. In any case,
the announced review of SPBA by the Swedish National Audit Office with a report
announced for December 2020 is most welcome, even if it does not go so far as
to analyze the type of problems considered here.
Example
#9 - Airliners’ and traffic’s control
Computer experts and researchers may think that the
examples up to now are "trivial" in the sense that they do not
represent the technical and logical complexities of modern HCI proper. They
were nevertheless enumerated because they transcend the processing of
visual-acoustic-haptic stimuli and "neurophysiological, instinctual,
behavioral automatisms’, and corresponding technical patterns" that are
often discussed in HCI-professional venues, and could be classified as cognitive ergonomics. In order to counter the accusation of
triviality in the above examples it is possible to adduce a more or most
complex representative example of this kind of "transcending" HCI: the so called Maneuvering Characteristics Augmentation System (MCAS),
which is a flight control
mode (software mode) that played in
the last years a role in airliner accidents, related to the Boeing 737 MAX groundings. This matter has already had enormous,
especially economic, consequences and therefore obviously has already been or
will be the object of extensive research. On the net there is available a
quite ambitious and detailed study by Krister
Renard written in Swedish with the title Boeing 737 Max – den sanna och (nästan) kompletta historien [Boeing 737 Max – the true and (almost) complete
story]. The Swedish text can be reasonably processed with Google translate, which in my experience has a good quality in
translations to/from English language. It integrates HCI with the whole
“system” reaching company boards. In a typically engineering, if not technocratic, mood he seems to believe that HCI implementation and
policies would improve if the power of decisions is not taken away from
engineers and given to graduate economists and lawyers.
Reflection on example
#9
It is symptomatic that HCI can be
seen as a heir of ergonomics and human factors, and it is worth to suspect (as I suggest in my text
on computerization as logic acrobatics) that it is the
absurdity of logic acrobatics that originates the doubtful pretense
of being able to create a hotchpotch or "combination of psychology,
sociology, engineering, biomechanics, industrial design, physiology,
anthropometry, interaction design, visual design, user experience, and user
interface design". That is, a mixture of disciplines and more or less
ephemeral "traditions" that also are seen as characterizing other
trendy fields or hypes such as AI and Design (cf. especially the latter's
sections on "Design disciplines" and "See
also"). The only research on this kind of problems
that has reached me and I have noticed up to now, is Alain Gras'
work in France, exemplified by two of the references, #2 and #3, in the French Wikipedia.
The only English translation I know is a condensed version of that above
#2 Le Pilote, le contrôleur et l'automate (Paris: Institut de Recherche et d'Information Socio-Economique IRIS Editions, 1990, ISBN
002-906860-06-93), co-authored by Gras, getting the title Faced with Automation: The Pilot, the Controller and the Engineer(trans.
Jill Lundsten, Paris: Publications de la Sorbonne, 1994, ISBN
2-8544-260X). It contains an (Introduction) For a socio-anthropology of
aeronautical technology, (Part I) Socio—historical construction of
the aeronautical system, (Part II) The attitudes and
representations of the airline pilots vis-à-vis innovation, (Part
III) Air traffic control and mutation, (Part IV) The
aircraft designers, and a (Conclusion) on Air logics and
ground logics, the new flight paradigm, sociology of appropriation, safety-the
core of civil aviation, and the return of the automation. I
take care to account for some details of this work, despite of its analysis
escaping the ambitions of my text, since its sheer layout illustrates a
contrast to many common HCI-approaches.
I guess that this approach may come to be criticized
for being only or too much “sociology” despite of sociology having been
mentioned as one of the legitimate watchwords in the combinatory definition of
ergonomics and human factors mentioned above. The complaining for “sociology”
is a remnant of the old positivistic attempt
to relieve engineers or administrators of the responsibility for the
consequences of their actions or use of their “tools”. Such positivism appears
also in Herbert Simon being considered a founding father of both artificial intelligence AI
and administrative theory. Ultimately, however, it is not a matter of sociology
but of a systemic view of HCI which is implied by and implies
the need for, and capacity of "combining" of all those scientific
"disciplines and more or less ephemeral traditions" represented by
the watchwords above. It is a need that is illustrated by all the examples
above, except for the last example #9above where the need is visible because
it is a question of life and death of people and business or a question of
politics and economics.
A defective understanding of these issues implies a
missed understanding of, for example, of the concept of “internal plan of
actions” advanced in Activity theory as mentioned in its article in Wikipedia. It defines
the internal plane of actions as
"[...] a concept developed in activity theory that refers to the
human ability to perform manipulations with an internal representation of
external objects before starting actions with these objects in
reality."
I think that was is missed, despite
the higher ambitions of activity theory relative to other HCI-approaches, is
that this is an age-old problem of simulation as related to
so-called reality. And this is valid also for the hype of virtual
reality. What was considered in the previous paragraph above has
already been problematized and considered in C.W. Churchman’s “Analysis of
the concept of simulation”. (in Hoggatt & Balderston,
eds. Symposium on Simulation Models, South-Western Publishing
Co., Cincinnati, Ohio, 1963, today more available in the author's summary
in The Systems Approach and its Enemies, 1979,
pp. 51-53). What I mean is that if one performed an analysis of the application
of “internal plane of action” in a non-trivial HCI-context one would find the
kind of complexities that appeared in the previous paragraph according to Le Pilote, le contrôleur et l'automate.
But, tragically this may
not be the case: no complexities, until further notice may be found by those
who believed in and implemented the Remote
air traffic control at the London City Airport, a news one could
believe it was fake before it was reported by Reuters on April 30, 2021 (not
April 1st) and PR-illustrated (on May
5th). Senior researchers may feel that they must leave it at that.
The need for combining knowledge disciplines beyond
technical logicalities becomes then visible in questions of life and death.
Logical minds who entrust powerful logical tools to simple-minded or credulous,
confiding, unsuspecting professionals or general public can pedagogically be
compared with naïve arms dealers putting machine-guns in the hands of
youngsters for their defense, assuming that they will not use them in ignored
or badly judged events. And also hoping that nobody succeeds in implementing a cost-benefit
analysis of the deal, which is assumed to be profitable. A test of the problem
of cost-benefits, that symptomatically no longer musters research enthusiasm,
can be intuited by reading e.g. “Measuring profitability impacts of information technology”
(pdf) in the ASIST 2003 meeting of American Society for
Information Systems and Technology, which recalls my earlier reference to
Ida Hoos’ critique. Logical minds’ assumptions in this context, suggest a
rough analogy to climate-activists’ reliance upon children’s leadership for implementing measures for
countering global warming.
The direction of ongoing HCI may be justified by the
idea that it is just a sort of ergonomic “handle” to allow the “handling” of a
powerful tool, disregarding the problematics of “tools” advanced earlier in
this text. After all, one could say, nobody objects against the modern “use”
of mathematics and related logic, or mathematical software, in the practice of modern science,
technology, and industry. There is also a rich supply of statistical software packages that can be
easily used and misused by professionals who have a very limited knowledge of
the presuppositions and pitfalls of the large-scale “production” of
mathematical and statistical results. Related problems in Sweden did motivate
the exemplary reaction of an experienced professional statistician, Olle Sjöström,
in a PhD dissertation (in Swedish, 1980, with English summary pp. 154-156)
on Swedish social statistics, ethics, policy and planning,
followed by a book (with English abstract, 2002) on Swedish history of statistics. The dissertation’s
main questions run very close to Edgar Dunn’s Social Information Processing and Statistical Systems (1974)
and are strategically presented as including
has the
growth in the “production” of statistical data led to improved information for
decision makers, better informed citizens and an increase of knowledge about
current social conditions? […]
a
discussion of human information processing in relation to statistical
data…Semantic problems which occur with such words as “data, “information”,
“knowledge, “decision under uncertainty” […]
It may look as if we are far from HCI but the point is
to remind the HCI tends to by its very nature to disregard its “production”,
the social context of its use, and consequences. In this context it is useful
to also recall a remarkable article and book chapter by Clifford Truesdell, described as American mathematician,
natural philosopher and historian of science, with the consciously chosen but
perhaps excessively provocative title “The computer: Ruin of science and threat to mankind”
(1980/1982). This provocation to which I have also referred in some
earlier papers, would have been expected to lead to many rebuttals, but
symptomatically I happen to find only one smug “quick rebuttal” of half a page that is not so quick,
being dated the year 2017. Paradoxically, it is a very self-confident rebuttal
that seems to me to confirm my views in this present text.
Humans are said to use software packages that allow
for a sort of HCI, human-centered acting on human-compatible computer software
that produces a reaction or result rather than a response, with an impact on
the same and other humans. As far as I humbly know about HCI, however, much of
it is at the level of so-called cognitive
ergonomics, which shares all doubts that there
must be about cognitive
science, which in turn shares all doubts about
what reason is or should be, especially since the philosopher
Immanuel Kant, and finally related to cognition or
“inquiry” in terms of the already mentioned design of inquiring systems.
Random examples include for instance recommendations for design of input to
computer communication networks along the lines of “Form design is a balance of science and art. It must be
functional and clear but also attractive”. Microsoft offers technical support
for creating forms that users complete or print in Word. That is, much of daily
HCI design would follow form design.
There are design enthusiasts who would even claim that high-flown “design
theory” or the “design perspective”, whatever theory, design and perspective
mean (if not the originally Kantian design-artistry, and the Nietzschean perspectivism), include the capability to guide interaction design. Another example would be the very intelligible and
recognizable exhortation to “intelligent response validation" that is “capable of detecting text input in
form fields to identify what is written and ask the user to correct (with a
maximum number of characters) the information if wrongly input. Please note
that the user is asked to “correct” what he writes, his supposedly wrong
informational input, and this is to be done within a maximum number of
characters; sometimes it is also specified that certain characters are not
allowed and that the whole must be done within a given time span, as it is the
case with internet-banks, for safety reasons.
These latter considerations recall
the case of personal shortcomings of a non negligible percent
of people, especially users, since clerks and other professionals would not at
all be employed and allowed to work with computers if they have a sort of dyslexia in
memorizing sequences of pressing keys of keyboards. Since a long time ago psychological testing has been used for selecting employees, e.g.
secretaries who need cognitive-ergonomic proficiency in matters like spelling,
language and numbers, or programmers and airplane pilots who need especially
logical-spatial proficiency. Those who have “keyboard dyslexia” or are
incapacitated by age may instead be allowed to survive so long as they
come-through in the increasingly computerized world, hoping for an option of
digitalized verbal input into whatever they do no longer understand. Nevertheless it is more than a question of cognition,
dyslexia or the like. Ultimately it may be a question of the limits of logicization or mathematization of reality as a consequence
of the computerization of digitalization of society that I have analyzed in another
article. It is a distortion of
reality that requires “distorted” minds. It does not fit minds of people who
are not mainly mathematically or logically oriented but, for instance, are
verbal and language oriented or, more sophisticatedly, oriented towards “thinking-sensation”
along the kind of types elaborated in analytical psychology. Those who have tried to live in computerized China
or to travel around in Europe and the world during the Covid-19
pandemic with the support of
special Covid
vaccination certificates, governmental health
certificates, matching passports, their QR-codes based on officially approved computerized registration forms, may already know what we are talking about. It
recalls the world of Franz
Kafka (1883-1924) who, as described in
Wikipedia “typically features isolated protagonists facing bizarre or
surrealistic predicaments and incomprehensible socio-bureaucratic powers”, which were
already presaged at his time. It resembles
the confusing, impossible “reality” of quantum
physics or of a sort of brave new world.
Complications can also be seen in
rich problematizations of technology and specialization (cf. HCI) along lines
that are not mine but can be sensed in e.g. Arnold Gehlen’s Die
Seele im technischen Zeitalter (1957, read 1986 in Italian, trans. A.
Burger) translated as “Man in the age of technology” (cf. other with same title),
esp. chaps.VIII and IX on automatism and personality.
Unperceived naiveté leads, however,
very far and very soon. For instance, a “grand old man” in the field, Gerhard Fischer wrote a very ambitious text brooding about the subject in terms of “user models” problematically and
brainly differentiated from “mental models”. It can lead to perceive that
“Traditionally, computer usage was
modeled as a human-computer dyad in which the two were connected by a narrow
explicit communication channel such as text-based terminals in
a time-sharing environment. The advent of more sophisticated interface
techniques, such as windows, menus, pointing devices, color, sound, and
touch-screens have widened this explicit communication channel.”
This means that information is conceived as in a communication
channel as I questioned in my doctoral dissertation at the beginning
of the seventies on Quality-control of information with
an unperceived relation to HCI including “forcing reality to fit the
model”, (chaps. 2.8 and 3.3). From such a conception it is easy to go further
arriving (p. 70) to the philosophically unperceived enormous question of the
“user” (whatever his definition among other neglected classes mentioned
above) saying “The right thing and the right time in the right way”,
which is a subtitle in Fischer's article. This expression should be put against
the background of my examples above in order to sense its inherent hubris. It
leads to Fischer’s further considerations such as
“user models are defined as models that systems have of users
that reside inside a computational environment. They should be differentiated
from mental models that users have of systems and tasks that reside in the
heads of users, in interactions with others”.
But is it “systems” or is it their designers who work
for managers who have models of users for the managers’ purposes of satisfying
(hopefully clients but mainly) shareholders, and then is it users who need to
have models of systems and their whole changing environment in order to know
whether and how to use them for their own purposes? And further:
“interaction between people and computers requires
essentially the same interpretive work that characterizes interaction between
people, but with fundamentally different resources available to the
participants. People make use of linguistic, nonverbal, and inferential
resources in finding the intelligibility of actions and events, which are in
most cases not available and not understandable by computers”.[…]
The ultimate objective of user modeling is that it has to be done for the
benefit of users. Past research has shown that there is often quite a
difference between modeling certain aspects of a user's work and behavior,
and applying this knowledge for the benefit of
the user”.
Yes, “linguistic, nonverbal, and inferential resources
in finding the intelligibility of actions and events, which are in most cases
not available and not understandable by computers”. But: “linguistic, nonverbal
and inferential resources” stand for the whole human mind, human science,
philosophy and culture which underlines and envelops the logic and mathematics
embodied in computers.
So, in trying to make it short, referring to the quotation above from Gerhard Fisher I put it in
the form of a question: is the ultimate objective of user modeling that it must
be done for the benefit of the user? It sounds as a sort of “Kantian
categorical imperative” but it may rather be a “paternal” diplomatic wishful
thinking, along with a patronizing computerization. It illustrates the main
message of my text: that simplicity hides complexity. What is going on is a
straitjacket-“logification” of the human mind along
the lines of “forcing reality to fit the model” where the model is logic. It
can contribute to people refusing to interact with or “use” computers, or in
their feeling distressed in the process of interaction, as in the case of
“keyboard dyslexia”. Or, worse: instead of refusing, people may no longer refuse
and not feel distressed, thanks to the human capability to adapt to the
environment. If computers cannot imitate humans as in the unbelievably and
symptomatically hyped naïve Turing
test, humans can imitate computers, as
several scholars like the above referenced Arnold Gehlen with regard to automatism and
personality, have already accounted for as a cultural phenomenon. A somewhat
less tragic alternative is that among those who do not refuse and do not feel
distressed, especially elders, there are people who are experiencing the Hawthorne effect. It
may be the case of lonely people, especially in care of elderly and disabled,
who are longing for human contact and love but feel temporarily happy in being
at least the object of attention and trials with the companionship and
assistance of dogs, computers and robots.
It is interesting to note that all this
can be conceptualized (not too far-fetched) as a mass-experiment that
is “democratically” imposed on a population in the name of science, namely
computer science and social engineering. In this respect it can be seen in the
light of Philip Zimbardo’s ethically problematic Lucifer
effect. In
my review (The
Lucifer effect and research ethics) of Zimbardo’s book, I
remark in a postscript that it was a prototype of mass-experiment,
according to Rutger Bregman’s Human Kind: A Hopeful History (A new history of
human nature), which contains an expanded severe criticism of Zimbardo’s Stanford Prison Experiment.
An analog historical wrongdoing in Sweden, on societal
scale is represented by the Vipeholm experiments,also
illustrated in the Swedish
public radio (1 nov. 2020) in
which disabled children were submitted to special diets in order to ascertain
the deleterious effect of sugars for caries. The facile justification of such
experiments in always based on that criticism is always easy ex-post, after the
fact but not ex-ante, before the event.
Symptoms of these problems appear already in mass
media and in social media, symptomatically still outside of, and unacknowledged
by the scientific literature. Two most touching examples and testimonies, more
touchingly and elegantly written than my nine examples above are available to
Swedish readers and to those who trust translations into English by Google translate. They are the political journalist
Anna-Lena Laurén’s ”App app app – det börjar likna
bolsjevism!” [“App
app app it's
starting to look like Bolshevism!”],
in the newspaper Dagens Nyheter, 14 october 2018, economic historian Ylva Hasselberg’s “Hejda
digitalbolsjevismen” [Stop the digital bolshevism],
in Dagens Nyheter, 23 december 2018d, and Arne Höök’s
”Frontsoldat
med omöjligt uppdrag” in Mitt i, 26 December 2020. They retell rhetorically their distressing experiences of forced
relation to computers, and associate it to the concept of arrogant neo-bolshevism, bolshevik technical rationalization zeal, a rebirth of forcing reality into
the model, which is also a source of the term political correctness. I ask
the reader of these lines of mine to allow my quoting a culturally justified,
illustrative digression in the field of literary essays; an example of how
“people” experience these problems. Excerpt from Bodil Malmsten’s book ” The
leopards coming from the north” (2009, in Swedish, my trans.,pp.
118-120):
Yesterday
was the day when my Internet service provider Orange
promises sending the Installer of Live
Box to get wireless internet here at the Atlantic Ocean
[where I live, in Finistère].
Orange.
It bodes ill, and indeed - the installer is one and a half hours late. When he
finally comes in an orange jacket with Orange logotype
I lock the door behind him and sit on the key until he - with phone support
from Bordeaux for there is installation firm's expert on Apple computers -
after four hours manages to install French wireless internet on my computers
where all menus are In Swedish. I force him to doublecheck everything before I
release him with the promise of a month's warranty.
Now
it was just to keep your breath and pray for my wireless life. Internet service
providers, are the same all over the world, throughout the universe, the ISPs
are a necessary evil and hell. The individual's vulnerability for the misuse of
the evil Internet suppliers, the expensive support numbers, the horrible music,
the disconnection, the constant problems.
One
day, I could neither dial nor receive on my phone, and I drive to the Orange store in the center of the city: I'm already angry
and, indeed, the phone problem is not Orange's problem, it is the
subcontractor's, the installer in the orange jacket with Orange's logo. A
malicious lady in orange in the Orange store refuses
to have to do with me, but since I refuse to leave the store before she calls
the subcontractor, she finally has to do it, forces herself to listen to
horrible music and push star and square until she just wants to die. When she
finally can talk after waiting at the queue, the representative of the
subcontractor replies that the only way to fix the line is that the customer
himself calls from his home phone.
Even
the malevolent Orange woman feels now stuck and tries
to explain that it is precisely because it is not possible to call from my home
phone that she calls from the Orange store at the
customer's - my - explicit request. "It is a special case," says Orange woman and stares unkindly at me. Everything is like a
movie of Jacques
Tati - not fun - and ends with the representative of the
subcontractor laying down the phone and the Orange woman saying the only thing
I can do is to drive home and pick up the live box and the phone and come back
to the Orange store so they can check whether my telephone is compatible with
the box. That my relatively new wireless phone could not be compatible with the
box is news, is something no one so far had said anything about, especially the
four-hour installer in the clothes with the Orange
logotype. I leave the Orange store, look the evil Orange woman in her eyes and say: "This is hell, c'est l'enfer, c'est l'enfer." [it is hell, it is hell.]
I
go to the neighbor who hates Orange over everything on earth, he has managed to
terminate his Orange subscription even though it is
impossible. The neighbor is, or course, willing to help me. But when he tries
to call the support number I have received from Orange, it turns out to be not
possible from his subscription with ISP Tele2. He still gets the issue solved,
and it is due to the neighbor never giving up, he never gives up, he does not
know how to give up. If the neighbor had been with Napoleon he would have won
the battle of Waterloo.
This leads further to the idea
of technological determinism, while my approach has been more of a philosophy of technology, which I cursorily approached in e.g. Trends in philosophy of technology. Today it gets hopelessly
fragmented in a mindblowing quagmire of
philosophy of computers, philosophy of artificial intelligence, and philosophy
of information, etc. Whatever is left of philosophy. It seems that democracy
under the guide of so-called science also tends to be perceived to become
bolshevism in the context of debates about global warming. Swedish readers can
again read about it in Lena Andersson’s “Demokratin är hotad om vetenskapen ensam ska styra klimatpolitiken” [Democracy is threatened if science alone is to control climate
policy], Dagens Nyheter, 28 sept. 2019, and
Hans Bergström’s “Att gömma sig bakom vetenskapen” [To hide oneself behind science], SvenskaDagbladet,
7 nov. 2020. I have already searched for
the explanation in my blog on Climate change and global warming, but ultimately in Information and theology as reduction of theology and science
to politics, when theology is not explained away by facile reference to the
need of a greater contact with the life
and the concrete body (theology
of the flesh), or accounting (but how?) for people’s values and valuations. When psychological and social science is not
reduced to politics it is because their politics is hidden by the recourse to a
superficial conception of the psyche.
It is the case of the study of Clifford Nass,
recognized authority in HCI. As expressed in Wikipedia "identifying a social theme in people's interaction
with computers, he was able to observe that humans project “agency” on
computers, and thus people will interact with computers in the same ways as
interacting with people". This idea was developed into The Media Equation, a book co-written
with Byron Reeves. I
have not noticed whether they consider that the "archetypal" case of
this phenomenon and its regrettable consequences in the history of computer
science and AI-mythology, is its appearance and naïve acceptance in the
popularization of the famous Turing Test. One main objection to
this kind of consideration should be and have been that it phenomenon
depends upon the level of maturity of the involved humans. The degree to which
humans feel (interaction with) computers as people is the degree to which
they project psychic content because
of it having been left unconscious in different degrees. If they have not yet
understood and felt what "people" and "humans" mean, they
will easily see and feel them to be equivalent to computers and the other way round.
What is then the role of sociology and
psychology in relation to politics and ethics as a theoretical background of
HCI? Which is the position of the HCI-relevant Activity
Theory and its Marxist roots
in Leontiev and Vygotski,
in this respect? What about the advertised ad-hoc categories of
“the individual, the object and the community”? How do they relate to the
interplay between work and capital, and to the “cultural sphere” with reference
to the broader fields of sociology, philosophy and theology that also are
culture in the lights of, say, a Dostoevsky and
his political reviewer N. Berdyaev? These are some of
the questions that were not considered in the otherwise
ambitious doctoral dissertation presented and disputed at my university in 1988
with the title of Work-oriented design of computer artefacts,
where work-oriented means very labor union oriented. Vygotski is
mentioned but the Czech Marxist philosopher Karel Kosik appears to be a main inspiration in the light of the
by now academically glorified Martin Heidegger whom I have dared to criticize
in several other contexts. A Marxist-Kosik framework or terminology was
used to meet, accommodate or neutralize my advisory objections about
considering the computer as a tool which was then renamed “artefact”. This lead
ultimately to postmodern reflections
and the hype of “design”, all criticized by Christopher Norris in a deconstructivist perspective
(which is not mine) in the book What’s Wrong with Postmodernism.
Together with the postmodern vagaries of corrupt Kantian aesthetics renamed “DESIGN” that
I denounced elsewhere both philosophically and theologically,
it all illustrates the hopeless vagaries of the intellect, the
impenetrable mindblowing intellectual
jungle of logical and ideological “impossible”
debates that can be fostered by the
misunderstood concept of reason, and awaiting a “Copernican revolution” consisting, I guess, of an (unlikely)
reconnection to theology.
QUAGMIRE OR
INTELLECTUAL JUNGLE
I note above the rise of Human compatibility, a new quagmire to be compared with e.g. Human-centered artificial intelligence earlier and elsewhere, both
of them implying conceptions of computer-human interaction. At the end of the
previous section I also notes the postmodern vagaries of corrupt
Kantian aesthetics renamed “DESIGN” that I denounced elsewhere both philosophically and theologically, it all illustrating the hopeless
vagaries of the intellect, the impenetrable mindblowing intellectual
jungle of logical and ideological “impossible”
debates that can be fostered by the
misunderstood concept of reason. I wish now to show how this reflects in the
ongoing attempts to develop the field of HCI as represented by the Association
for Computing Machinery’s (ACM) Special Interest Group (SIG) for Computer-Human
Interaction (SIGCHI)
with its annual international SIGCHI past
and planned
conferences. In the present context I will not yet adduce the corresponding
organ of the International Federation for Information Processing’s (IFIP)
Technical Committee number 13 on Human-Computer Interaction (IFIP-TC13)
with its working
groups and planned international “INTERACT” biannual
international conferences.
To illustrate the ambitions of next online virtual conference CHI-2021, in May 8-13,
2021, anticipating more than 3000 paper submissions, there are no less than 15
subcommittees which may be considered by some as “disciplines”. Prospective authors have been asked to suggest a
subcommittee for their submissions, but it is not clear to which
subcommittee the present paper of mine could have been sent except possibly for
the mild User Experience and Usability,
if it is supposed that “usability” is a reliable concept, allowing for a sort
of operational definition, and encompassing my nine given very concrete
examples:
1.
User
Experience and Usability
2.
Specific
Applications Areas
3.
Learning,
Education, and Families
4.
Interaction
Beyond the Individual
8.
Health
10.
Design
11.
Interaction
Techniques, Devices, and Modalities
12.
Understanding
People: Theory, Concepts, Methods
13.
Engineering
Interactive Systems and Technologies
14.
Critical
and Sustainable Computing
I already pointed out above that it is the absurdity of simple logic thinking that originates the
doubtful pretense of ergonomics
evolving into human factors, being able to create a hotchpotch of,
or claiming to draw on many disciplines in its study of humans and their environments,
including anthropometry, biomechanics, mechanical engineering, industrial engineering, industrial design, information design, kinesiology, physiology, cognitive psychology, industrial
and organizational psychology, and space psychology. It is weird to compare
this list with the list above. That is, a mixture of disciplines with more or
less ephemeral "traditions" that also happen to characterize other
trendy fields or hypes such as artificial intelligence (AI) and general “Design”.
Disciplines in research contexts, however, used to be identified as such if
they deserved to be considered as academic disciplines. Such strict and historical view has already been
shaken with the result that it is claimed that academia is being replaced by
different modes of production of knowledge including “post-academic science”
while universities
themselves gradually lose the “idea of university”
compared with a research lab in industry. It is this that comes in display in
the list of “subcommittees” for HCI, a list that may even be different for
every recurring annual meeting. Anybody who has been working or teaching in
whatever “area” for 20-30 years may be considered as expert.
1. Describing
and reflecting on existing philosophically-informed HCI research.
2. Identifying
emerging phenomena, issues, and challenges in HCI, addressing which would
benefit from engaging with philosophy.
3. Considering
how we can practically learn and “do” philosophy. For instance, how might we
explicitly meld traditional HCI methodology stemming from disciplines like
psychology with philosophy?
4. Asking
what philosophy means for understanding stakeholders and for design of
interactive systems.
5. Developing
how can we make philosophy in HCI more accessible.
6. Outlining
an agenda (or agendas) for philosophically informed HCI.
Nevertheless, I have found that none in neither the list
of workshop’s accepted Position Papers for this
workshop, nor in the list of the
conference’s Extended Abstracts of the 2019 CHI conference on Human Factors in
Computing Systems seems by far to cover the problems and content of the present
paper of mine, if it is “philosophy” or whatever. There are, however, workshop position papers just mentioning
Kant, phenomenology and Heidegger. Ludwik Fleck is also referred
to, a physician and biologist whose philosophizing developed the concepts
of e.g. thought style, logology, (science of science), and thought collective (cf. collective
consciousness and culture, avoiding the
quagmire of Spengler’s distinction
between culture and civilization), about the same time as Carl Jung
wrote about the ignored parallel type psychology and collective unconscious. A feministic
approach to the workshop on philosophy refers
mainly Critical theory with unknown connection HCI, and it is
not mentioned by other feminist such as Sherry Turkle above.
There is a workshop position paper on Chinese
philosophy with problematic unknown relation to
computers but I guess that a connection between Chinese and Western philosophy
related to information systems for the purpose of HCI-research is to be found
in my paper on Chinese information systems? East and West. The conference paper in 2019 CHI, Models of the Mind reveals what happens also to be considered as
“philosophy” when stating that it draws on “philosophies of embodied,
distributed & extend [sic] cognition, claiming the mind is readable from
sensors worn on the body. Another conference paper on Failing with style (cf. the above
reference to Churchman’s “The success of failure”) may claim
that “failure is a common artefact of challenging experiences, a fact of life
for interactive systems but also a resource for aesthetic and improvisational
performance”. One can wonder whether that is applicable to at least some of my
examples. All these implied difficulties of loose philosophical connections to
HCI tend to be comfortably submerged in HCI-neologisms and in occasional claims
that its research has perceived the problems exposed in this article, at least
“aesthetically”, but not yet researched them.
In other words, the late “expansion” of HCI research
beyond the original, supposedly simple and naïve strict cognitive ergonomics,
whatever cognitive means or should
mean, does not seem to address any of the problems of this paper. It is
possible that HCI is expanding everywhere, or anywhere, and perhaps nowhere. It
reminds me of my quotation of Jan Brouwer in the context
of computers seen as embodiment of logic and mathematics:
Every branch of science
will therefore run into deeper trouble; when it climbs too high it is almost
completely shrouded in even greater isolation, where the remembered results of
that science take on an independent existence. The "foundations" of
this branch of science are investigated, and that soon becomes a new branch of
science. One then begins to search for the foundations of science in general
and knocks up some "theory of knowledge".
I submitted the
above text of this section on Quagmire
and intellectual jungle to a senior researcher and professor who dedicated
most of his life to HCI. Here follow the comments I received in two consecutive
mails:
First of all, I think your examples are very well presented and point to
REALLY important issues. The shameless abuse of power by companies and lack of
basic empathy to people (those who they think can be ignored), are
discriminating and disgusting.
What I do not completely agree with, is that HCI is not mostly about
cognitive ergonomics. It used to be the case, but it is changing. The scope of
HCI is expanding, and in 2019 there was even a workshop on philosophy at the
CHI conference […].
This year the CHI conference introduced a new sub-committee,
"critical and sustainable computing", which deals with some relevant
issues: https://chi2021.acm.org/for-authors/presenting/papers/selecting-a-subcommittee#Critical-and-Sustainable-Computing.
I think your personal stories of frustrating power struggles with the
faceless inhumane entities behind "human-computer interaction" resonate well with some of the themes covered by the
subcommittee.
I agree with you that conceptually HCI now is a mess. It is a mixture of
different agendas, approaches, etc. It used to be conceptually consistent, but
not any longer. And it does not properly build on humankind's intellectual
history. I disagree that HCI hasn't studied the issues you are highlighting in
your examples. Many of them (even though probably not all of them) have been
actually studied, or at least pointed out as relevant research issues.
So, I think the issues you mention are being addressed, but often in an
incomplete, naive, and inconsistent way.
Why? I think there are many reasons. The current fragmentation of
knowledge is probably one of them, and another one, in my view, is that HCI
research is trying to address lots of different assorted, often concrete and
practical, needs in knowledge about human uses of digital technology. These
requests for HCI-type studies are outside researchers' control (they just see
that one can get project grants for this and that).
What
can be done about it? Beats me...
What I perceive as a general intellectual quagmire is illustrated by two
books that are interrelated through one of their authors, and deal with Primitive Interaction Design (2020, 133 pp. originally priced at US$ 170, sic) and The Psychosocial Reality of Digital Travel (2022). The latter I have already commented in another
paper less related to HCI, with emphasis on the concept of “reality”. With
regard here to the first mentioned book its preface states (pp. viii-ix, my
italics) that
Tangible interaction,
unconsciously executed and informed by peripheral information restores the
primacy of action and re-integrates the
mind and the body. The technology
then disappears from perception in use. That means that products/artefacts
have to be designed for human beings,
not for users or customers. […] To be
pleasant and invigorating, human life should be free of the need to always be
conscious of the environment in which it exists. […] …our thinking of design
and information-based society should adapt by using more a more universal
approach and aspects of human consciousness/unconsciousness in a new,
“primitive” coexistence with modern information technology.
Such a program already contradicts the position of my own present text
inasmuch it is dedicated to the study of the interaction between the human
beings differentiated among users or customers, decision-makers, and designers.
The program seems instead to be more directed towards the idea of interaction
as all-encompassing aesthetic entertainment as suggested by
the second book mentioned above, on digital travel by the use of virtual
reality. Entertainment is necessarily a means both aesthetics and consumption
despite any claims of attaining morality through an integration of mind
(whatever it is) and body, or as I identified as the “flesh” in a study
of phenomenological ethics. But the book claims more than so (p.15):
To find inspiration for
the new view of design presented in this book, and to change attitudes about
designing. We have looked beyond conventional design to the methodological
playgrounds of anthropology, mythology, theology, science, ethics and art. […]
A consideration of the spiritual dimension, and of myth, of emptiness and of
the unconscious should come as a revelation to the profession […].
I find that it claims too much, while not mentioning philosophy. Being
aesthetics, it also incurs in its neglected Kantian relation to reason and
ethics under the label of “design”, that is covered in two essays of mine about computers
as embodied mathematics and logic, and especially the one about computerization. The reader of
these lines who wishes to spare the effort of reading and understanding the
rest of this section may go over to study my latter reference dealing with
computerization. The whole book on Primitive Interaction Design considered here
is in my view a misunderstanding, a failed attempt to see “design” as a way to
salvation of humanity by reducing all philosophy to phenomenology and avoiding
the mention of Christianism. In fact, the book reminds (p. 33) that in the
pursuit of science
God was declared dead
and religion relegated to one day a week, at best. The separation of our minds
from our bodies, our reason from our emotions, was complete. […] However and surprisingly, this most technological and
abstract of inventions [computer technology and associated communications
capabilities] came to provide the means for a reintegration of being and doing.
Virtual reality […].
The book does (not) so by recurring mainly to unbounded wholesale
references to authors related to phenomenology, which
symptomatically overflows into the Wikipedia’s “See
also” related to phenomenology. This adduction of phenomenology becomes
possible by means of the adoption of an approach that has been called Fact-Nets or Leibnizian Inquiring Systems, mentioned in my survey on Information and Debate.
Skill in construction of fact-nets allows, for e.g. an introduction to
“the main dimensions of the mind” based on interaction between the Conscious and the Unconscious (p. 37f.)
followed by consideration of emotion)
without a single reference to the otherwise ambitious bibliographies at the end
of each chapter of the book. Otherwise it is this
approach to fact-nets that allows for an apparently enormous coverage by the
book’s total of 133 pages. References to Carl Jung’s collective unconsciousness (unconscious), immediately followed by Buddhistic thought later in the book (p. 77ff.), however,
are not applied to explain Jung’s
conception and the earlier mentioned interaction between the conscious and the
collective unconscious. Neither are they applied to later (p. 97) tools that embody collective
unconsciousness. Not to mention the problem of understanding what consciousness is, as evidenced by repeated frequent confused claims
from several quarters, such as Physicist claims to have solved the mystery of
consciousness (August 14,
2022). Forget Erich
Neumann’s The Origins and History of Consciousness (1970/1949). I
think that many readers will get impressed by the our
considered book with its ambitious coverage of apparently disparate conceptual
frameworks, in what can be perceived as a somewhat rhetorically bloated style,
as exemplified by the label of a figure (p. 123): Conceptual spaces for a blended morphogenetic prototyping environment.
I must confess that I perceive this kind of approach as “mind-blowing”
with the risk that it says more about my mind than about the approach. This is
aggravated by the fact that the philosophy and psychology refer to Buddhistic
thought without any reference to neither Christianism nor to the differences
and relations between western and eastern thought, such as found in Carl Jung’s
thought. Not to mention the pitfalls in western people trying to understand in
depth eastern thought when they barely understand their own Christian culture.
The book announces already in its preface (p. viii, my italics):
Tangible interaction,
unconsciously executed and informed by peripheral information, restores the
primacy of action and re-integrates the mind and the body. The technology then disappears from perception in use. That means
that products/artefacts have to be designed for human beings, not users or
customers.
[…] To be pleasant and
invigorating, human life should be free of the need to always be conscious of
the environment in which it exists.
Disregarding what an undefined “environment” is supposed to be, my
conclusion is that paradoxically this is what happened with the authors. They
do not define themselves as users or customers but they are either undefined
designers or most probably see themselves as just human beings. Technology has disappeared from their
perception in use as HCI designers and authors while paradoxically it
permeates the whole book whose structure of facts-nets embodies the essence of
technology. It is my perception that scientists and technicians who deal with
computers such as HCI are proficient in logical-mathematical thinking that
expresses itself in fact-nets. Analytical
psychology would hypothesize that while they consciously deny
and repress their inclination to logical-mathematical thinking they are
overpowered by indiscriminate feelings to unconsciously express it in their
professional design and authoring activities for which they were employed in
academia, in the first place.
I dare to leave it at that, asking what is the place
of formal sciences embodied in computers, as related to other sciences, as
suggested in my other works on mathematics, digitalization and theology.
Long after I had written and published this paper I wrote another weird text
with the title Logic as rape vs. Truth and Love, and a bit later I realized that this present paper
advanced examples of mental rapes. In
the meantime I had experienced and witnessed a number
of additional better examples of HCI breakdowns, which would have justified a
rework and lengthening of this paper to its double or triple size, if I had
been young.
The question is more than a fragmentation of knowledge mentioned above. It is also the original
and basic fragmentation between formal logical-mathematical knowledge furthered
by computers, and other knowledge, as well as what knowledge really is or
should be. The question is then whether the computer is forcing human psyche to
become a “brain” using (in Brouwer’s sense) only a small “logical” part of
itself in reflection and communications between people and with cultural
history. And whether this can or cannot be obviated by means of reforms of
logic and by eclectic patchworks of logically structured so-called user and
mental models that imply an imperceptible gradual impoverishment of dialog and
culture. Against this background I have refrained from the impossible task of
reviewing all the canonic literature of HCI beyond samples from some of its
luminaries as exposed, for instance, in the “Footnotes” and “Further reading”, in Wikipedia. It would have incurred
in the problems of “Debate” as I depict them in the context of
academic cronyism and socialization.
For the rest I am obviously aware that the interaction
with computers brings innumerable advantages, and that my text may be
interpreted as being too negative, not acknowledging the advantages of a
developing technology with only transitional childhood-diseases. Those who
interpret and feel so, usually neglect that advantages and disadvantages, or
benefits and costs, are concepts of 300 years-old utilitarianism and therefore they also neglect
the criticism that has been directed at it. For the rest, I will not go so far
as to defend my text only by referring to the Faustian
bargain and its history. And I do not believe that it is neo-luddite technophobia
to ask whether all latest technical innovations always mean better future for
everybody, equaled to that more is achieved faster and cheaper for everybody,
without knowing where and how it will end if not in a paradise. I
can only affirm that it is my basic positiveness towards technology that
motivated me to five years of hard university studies allowing me to start a
career as electronic engineer and to study most of the time technical matters
at institutes of technology or what today is often called technical
universities. And to deal with basically technical disciplines. It is only age
(now 83) that prevents me from having the time and energy for further studies
which would extend my explanatory thoughts and my writing beyond what I already
wrote in Trends in the philosophy of technology and Information and theology, and
extend my readings beyond Theology and technology.
Let me terminate with quoting from my earlier essay on
the drive towards computerization, as it impacts the
discussion of HCI. In July 2021, the Reuters news agency announced a
major "private ransomware-as-a-service" (RaaS) attack by REvil on the U.S. tech provider Kaseya,
which forced the Swedish Coop grocery store chain to close all 800 of its
stores for about a week: I wrote:
The late
and perhaps ultimate consequence of short-circuiting the human and social
element in the increasingly inclusive logical processes is the expanding
phenomenon of Ransomeware and in
particular of Ryuk under the general labels of Cyberattacks and Computer
security. I do not know of any ex-post "after-the-event" corrective except smart back-ups and the
hopeless hodgepodge of the indefinitely expanding "security industry"
including FDIR - Forensic Digital Incident Response, to be considered as a sub-field of FDI - Fault detection and isolation. As ex-ante preventive measure I can only naively
idealistically think of my dissertation on Quality-control of information, and the
related summary in the book (in Swedish) on Systems development and
rule of law.
In order to avoid that HCI will mainly contribute to
the growth of discontent at the edge of HCI-luddism,
allurement for criminal organizations, and consequent challenges to a growing
hodgepodge of "security industries" leading to an analogue of a (cyber)
"arms
race", we have to acknowledge that knowledge cannot
be reduced to computerized logic or so-called Leibnizian
inquiring systems. One cannot by logical means prevent the
breakdown of logical chains and networks. I can now, in an apparently utopian
attitude, add that what has been expressed in the present text is the ultimate ex-ante preventive measure.
Let me terminate by also reminding that
the ongoing computerization of digitalization of society means that an
increasing number of elders get handicapped by not being able to necessary
daily tasks requiring digital “dexterity”. The first consequence of beginning dementia
besides failures of memory is the weakening of logical thinking, which is
hidden behind vague expressions such as Experiencing
memory loss - poor judgment - and confusion, Trouble handling money responsibly and paying bills, or Taking longer to complete normal daily tasks.
My experience is that, among all this, the weak link
is elementary logical thinking. A man I know has a neighbor that at the age of
75-76 after successfully having used regular mail, cellular phone and computer
suddenly could no longer pay his bills that he had processed manually by
regular mail. He hoped that he could put all business letters and invoices in
an envelope and mail them to his bank that would take care of the rest. Since
his closest relatives live in other cities he began to
rely on his neighbor who could take photos of all correspondence and send it
digitally to the man’s son who succeeded in arranging being a proxy with an
authorized own digital contact with his father’s bank and pay his bills
digitally. All this in expectation that the father in the future will move to
live near the son. All this affecting also those who are not yet victims of
beginning dementia recalls what I did already write above and repeat now as a
description of what is like an ongoing societal experiment right now:
Those who have “keyboard dyslexia” or are incapacitated by age may
instead be allowed to survive so long as they come-through in the increasingly
computerized world, hoping for an option of digitalized verbal input into
whatever they do no longer understand. Nevertheless it
is more than a question of cognition, dyslexia or the like. Ultimately it may
be a question of the limits of logicization or
mathematization of reality as a consequence of the computerization of
digitalization of society that I have analyzed in another article.
By the end of the year 2024 I had got personal
experience and knowledge of about 35 cases of problems in human-computer
interaction that would have completed and deeply illustrated the cases advanced
in this paper. Long after the rest of this paper was written, in October 2025 I
had one case of suspected preparation of bank fraud that succeeded in
convincing me to include a tentative
description of it in my blog. It exemplifies the absurd effort that
would have been required to report the other earlier 35 problems and more since
the end of the past year 2024.
I happened to
describe a few additional simple cases in the introduction to an essay of mine,
which I had not foreseen would be among the most downloaded and read, since it
deals with personal experiences of problems of medical safety
as patient, outside my professional area of
interest. They remind that the issue of security has to be expanded by
including (the security of) privacy as best exemplified by the increasing
discussion of privacy, considered for the time being (April 2023) in
Wikipedia’s (provisory) title Privacy concerns with Facebook.
Other considerations are presented in another essay of mine on the issue of computerization of society.
The problems cannot be resolved by logical computer
means, even as formalized in law, for the very same reasons that computer
security cannot be resolved by logical means. They increase my conviction about
the character and seriousness of this situation and its development. It
requires further attention by researchers who are not limited in time and
energy as I feel at my present age (87). I hope that they feel encouraged by
what I already wrote. The more so with the advent of anonymous “interactions” with
or rather through the latest versions of so-called Artificial General
Intelligence with the warnings for its “existential risks” that I partly
consider in the last section or epilogue of my above mentioned essay on the
computerization of society, and finally in my text on Artificial General Intelligence and ChatGPT. In that text, as well
in a section of Reason and Gender, becomes gradually more
visible that human-machine interaction implies also a progressive substitution
of humans by computing machines. This to the point that a fraction of humans
who keep really interacting with and
not only using with computers
gradually bypasses the thoughts and emotions of an increasing number of humans
who do not affect the contents and the operations of the interacted computers.
This up to the point that HCI keeps losing a part of its identity, the meaning
of interaction, because the number of human really participating in it decreases.
The interaction takes place between a gradually decreasing number of active participants while the number of
passive (including unemployed!) “users” (“users” or consumers of the effects of
a few others’ production of HCI-features, capabilities and presuppositions),
increases.
The follow-up of all this and the updating of this
paper would then require continuous research of a group dedicated to the issue.
It is indeed the follow-up of a real-time irreversible process of society’s
increasing computerization, which amounts to an ongoing irreversible
“experiment” with unforeseen long-term consequences. An example may suffice. I
know the case of an aged lady who in year 2016, without foreseeing a cascade of
successive versions and other iPhone generations and models, had downloaded on
her first
generation’s iPhone SE (Model MLM62KS/A, Series
F19RXJ8NH2XN, 64 GB memory) the SL-app
(see also here and
here) of
“Storstockholms Lokaltrafik”
(whose English website-link did not work on e.g. October 18 2024) for planning
a trip with the local public transportation systems. Despite not buying new
generations or models, but guaranteeing automatic downloads of updated versions
of the app’s operating system, the lady was able to use her iPhone until year
2024 when it suddenly began showing on the screen that “Something was
failing/was wrong). She then phoned to the SL-customer service and was informed
that “Yes, other clients had already perceived problems with the app”. It
seemed to mean that some bug was to be repaired in the close future, but this
appeared not to be the case.
The case appears to be that without any warning some
of the latest updates of the app presupposed the later updates of the operating
system, which presupposed the structure of the later versions
of the iPhone SE, all the while various models are discontinued
and unsupported, discontinued and supported, or unsupported after a “last OS”.
This series of events is analog to why certain services such as bank-payments
in updates of browsers like Safari, with no previous warnings, suddenly are not
possible in cellular phones, iPads and computers, which avoid continuous
updates of operating systems because of incompatibilities with other installed
programs. This is also the reason of why many documents in digital memories
will not be readable in the future, all the while services like the archive.org
were attacked with DDOS in
October 2024.