Can a technology be true?

Dave Cormier is a wonderfully sideways-thinking writer, such as in this recent discussion of the myth of learning styles. Dave’s post is not mainly about learning style theories, as such, but the nature and value of myth. As he puts it, myth is “a way we confront uncertainty” and the act of learning with others is, and must be, filled with uncertainty.

impression of someone with many learning stylesThe fact that stuff doesn’t have to be true to be useful plays an important role in my latest book, too, and I have an explanation for that. The way I see it is that learning style theories are (not metaphorically but actually) technologies, that orchestrate observations about differences in ways people learn, to attempt to explain and predict differences in the effects of different methods of teaching. Most importantly, they are generative: they say how things should and shouldn’t be done. As such, they are components that we can assemble with other technologies that help people to learn. In fact, that is the only way they can be used: they make no sense without an instantiation. What matters is therefore not whether they make sense, but whether they can play a useful role in the whole assembly. Truth or falsehood doesn’t come into it, any more than, except metaphorically, it does for a computer or a car (is a computer true?). It is true that, if the phenomena that you are orchestrating happen to be the findings and predictions of science (or logic, for that matter) then how they are used often does matter. If you are building a bridge then your really want your calculations about stresses and loads to be pretty much correct. On the other hand, people built bridges long before such calculations were possible. Similarly, bows and arrows evolved to be highly optimized – as good as or better than modern engineering could produce – despite false causal reasoning.  Learning styles are the same. You can use any number of objectively false or radically incomplete theories (and, given the many scores of such theories that have been developed, most of them are pretty much guaranteed to be one or both) but they can still result in better teaching.

For all that the whole is the only thing that really matters, sometimes the parts can be be positively harmful, to the point that they may render the whole harmful too. For instance, a pedagogy that involves physical violence or that uses threats/rewards of any kind (grades, say), will, at best, make it considerably harder to make the whole assembly work well. As Dave mentions, the same is true of telling people that they have a particular learning style. As long as you are just using the things to help to design or enact better learning experiences then they are quite harmless and might even be useful but, as soon as you tell learners they have a learning style then you have a whole lot of fixing to do.

If you are going to try to build a learning activity out of harmful parts then there must be other parts of the assembly that counter the harm. This is not unusual. The same is true of most if not all technologies. As Virilio put it, “when you invent the ship, you invent the shipwreck”. It’s the Faustian bargain that Postman spoke of: solving problems with a technology almost invariably creates new problems to be solved. This is part of the dynamic the leads to complexity in any technological system, from a jet engine to a bureaucracy. Technologies evolve to become more complex (partly) because we create counter-technologies to deal with the harm caused by them. You can take the bugs out of the machine, but the machine may, in assembly with others, itself be a bug, so the other parts must compensate for its limitations. It’s a dynamic process of reaching a metastable but never final state.

Unlike bows and arrows, there is no useful predictive science of teaching, though teaching can use scientific findings as parts of its assembly (at the very least because there are sciences of learning), just as there is no useful predictive science of art, though we can use scientific findings when making it. In both activities, we can also use stories, inventions, beliefs, values, and many other elements that have nothing to do with science or its findings. It can be done ‘badly’, in the sense of not conforming to whatever standards of perfection apply to any given technique that is part of the assembly, and it may still be a work of genius. What matters is whether the whole works out well.

At a more fundamental level, there can be no useful science of teaching (or of art) because the whole is non-ergodic. The number of possible states that could be visited vastly outnumber the number of states that can be visited by many, many orders of magnitude. Even if the universe were to continue for a trillion times the billions of years that it has already existed and it were a trillion times the size it seems to be now, they would almost certainly never repeat. What matters are the many, many acts of creation (including those of each individual learner) that constitute the whole.  And the whole constantly evolves, each part building on, interacting with, incorporating, or replacing what came before, creating both path dependencies and new adjacent possible empty niches that deform the evolutionary landscape for everything in it. This is, in fact, one of the reasons that learning style theories are so hard to validate. There are innumerable other parts of the assembly that matter, most of which depend on the soft technique of those creating or enacting them that varies every time, just as you have probably never written your signature in precisely the same way twice. The implementation of different ways of teaching according to assumed learning styles can be done better or worse, too, so the chances of finding consistent effects are very limited. Even if any are found in a limited set of use cases (say, memorizing facts for a SAT), they cannot usefully predict future effects for any other use case. In fact, even if there were statistically significant effects across multiple contexts it would tell us little or nothing of value for this inherently novel context. However, like almost all attempts to research whether students, on average, learn better with or without [insert technology of interest here], on average there will most likely be no significant difference, because so many other technologies matter as much or more. There is no useful predictive science of teaching, because teaching is an assembly of  technologies, and not only does the technique of an individual teacher matter, but also the soft technique of potentially thousands of other individuals who made contributions to the whole. It’s uncertain, and so we need myths to help make sense of our particular, never-to-be-repeated context. Truth doesn’t come into it.

Some thoughts for Ada Lovelace Day

This Scientific American article tells the tale of one of the genesis stories of complexity science, this one from 1952, describing what, until relatively recently, was known as the Fermi-Pasta-Ulam (FPU) problem (or ‘paradox’, though it is not in fact a paradox). It is now more commonly known as the Fermi-Pasta-Ulam-Tsingdou (FPUT) problem, in recognition of the fact that it was only discovered thanks to the extraordinary work of Mary Tsingou, who wrote the programs that revealed what, to Fermi, Pasta, and Ulam, was a very unexpected result. 

The team was attempting to simulate what happens to energy as it moves around atoms connected by chemical bonds. This is a classic non-linear problem that cannot be observed directly, and that cannot be solved by conventional reductive means (notwithstanding recent work that reveals statistical patterns in complex systems like urban travel patterns). It has to be implemented as a simulation in order to see what happens. Fermi, Pasta, and Ulam thought that, with enough iterations, it would reveal itself to be ergodic: that, given long enough, every state of a given energy of the system would be visited an equal number of times. Instead, thanks to Mary Tsingou’s work, they found that it was non-ergodic. Weird stuff happened, that could not be predicted. It was chaotic.

The discovery was, in fact, accidental. Initial results had shown the expected regularities then, one day, they left the program running for longer than usual and, instead of the recurring periodic patterns seen initially, it suddenly went haywire. It wasn’t a bug in the code. It was a phase transition, perhaps the first unequivocal demonstration of deterministic chaos. Though Fermi died and the paper was not actually published until nearly a decade later, it is hard to understate the importance of this ‘accidental’ discovery that deterministic systems are not necessarily ergodic. As Stuart Kauffman puts it, ‘non-ergodicity gives us history‘. Weather is non-ergodic. Evolution is non-ergodic. Learning is non-ergodic. We are non-ergodic. The universe is non-ergodic. Though there are other strands to the story that predate this work, more than anything else this marks the birth of a whole new kind of science – the science of complexity – that seeks to deal with the 90% or more of phenomena that matter to us, and that reductive science cannot begin to handle. 

Here’s a bit of Tsingou’s work on the program, written for the MANIAC computer:

Mary Tsingou's original algorithm design, drawn in freehand

It was not until 2008 that Tsingou’s contribution was fully recognized. In the original paper she was thanked in a footnote but not acknowledged as a co-author. It is possible that, had it been published right away she might have received proper credit. However, it is at least as possible that she might not. The reasons for this are a mix of endemic sexism, and (relatedly) the low esteem accorded to computation at the time.

The relationship between these two factors runs deep.  Historically, the word ‘computer’ originally referred to a job title.  As scientists in the 19th Century amassed vast amounts of data that needed processing, there was far too much for an individual to handle. They figured out that tasks could be broken up into smaller pieces and farmed out in parallel to humans who could do the necessary rote arithmetic.  Because women were much cheaper to hire, and computing was seen as a relatively unskilled (albeit very gruelling and cognitively demanding) role, computing therefore became a predominantly female occupation. From the 19th Century onwards into the mid 20th Century, all-women teams worked on astronomical data, artillery trajectories, and similar tasks, often performing extremely complex mathematical calculations requiring great precision and endurance, always for far less pay than they deserved or that a man would receive. Computers were victims of systematic gender discrimination from the very beginning. 

The FPUT problem, however, is one that doesn’t lend itself to chunking and parallel computation: the output of one iteration of the computation is needed before you can calculate the next. Farming it out to human computers simply wouldn’t work. For work of this kind, you have to have a machine or it would take decades to come up with a solution.

In the first decade or so after digital computers were invented significant mathematical skill was needed to operate them. Because of their existing exploitation as human computers, there was, luckily enough, a large workforce of women with advanced math skills whose manual work was being obsoleted at the same time, so women played a significant role in the dawn of the industry. Mary Tsingou was not alone in making great contributions to the field.

By the 1970s that had changed a lot, not in a good way, but numbers slowly grew again until around the mid-1980s (a terrible decade in so many ways) when things abruptly changed for the worse.

graph showing the huge drop in women in IT from the 1980s onwards

Whether this was due to armies of parents buying PCs for their (male) children thanks to aggressive marketing to that sector, or highly selective media coverage, or the increasing recognition of the value of computing skills in the job market reinforcing traditional gender disparities, or something else entirely (it is in fact complex, with vast self-reinforcing feedback loops all the way down the line), the end result was a massive fall in women in the field. Today, less than 17% of students of computer science are women, while the representation of women in most other scientific and technical fields has grown considerably.

There’s a weirder problem at work here, though, because (roughly – this is an educated guess) less than 1% of computer science graduates ever wind up doing any computer science, unless they choose a career in academia (in which case the figure rises to very low single figures), and very few of them ever do more mathematics than an average greengrocer. What we teach in universities has wildly diverged from the skills that are actually needed in most computing occupations at an even sharper rate than the decline of women in the trade. We continue to teach it in ways that would have made sense in the 1950s, when it could not be done without a deep understanding of mathematics and the science behind digital computation, even though neither of these skills has much if any use at all for more than a minute fraction of our students when they get out into the real world. Sure, we have broadened our curriculum to include many other aspects of the field, but we don’t let students study them unless they also learn the (largely unnecessary in most occupations) science and math (a subject that suffers even lower rates of non-male participation than computing). Thinking of modern computing as a branch of mathematics is a bit like treating poetry as a branch of linguistics or grammar, and thinking of modern computing as a science is a bit like treating painting as a branch of chemistry. It’s not so much that women have left computing but that computing – as a taught subject – has left women. 

Computing professionals are creative problem solvers, designers, architects, managers, musicians, writers, networkers, business people, artists, social organizers, builders, makers, teachers, or dreamers. The main thing that they share in common is that they work with computers. Some of them are programmers. A few (mostly those involved in designing machines and compilers) do real computer science. A few more do math, though rarely at more than middle school level, unless they are working on the cutting edge of a few areas like graphics, AI, or data science (in which case the libraries etc that would render it unnecessary have not yet been invented).  The vast majority of computing professionals are using the outputs of this small elite’s work, not reinventing it. It it not surprising that there is enormous diversity in the field of computing because computers are universal machines, universal media, and universal environments, so they encompass the bulk of human endeavour. That’s what makes them so much fun. If you are a computing professional you can work with anyone, and you can get involved in anything that involves computers, which is to say almost everything. And they are quite interesting in and of themselves, partly because they straddle so many boundaries, and ideas and tools from one area can spark ideas and spawn tools in another.

If you consider the uses of computer applications in many fields, from architecture or design to medicine or media to art or music, there is a far more equal gender distribution. Computing is embedded almost everywhere, and it mostly demands very different skills in each of its uses. There are some consistent gaps that computing students could fill or, better, that computing profs could teach in the context they are used. Better use could be made of computers across the board with just a little programming or other technical skills. Unfortunately, those who create, maintain, and manage computers and their applications tend to mainly come out of computer science programs (at least in North America and some other parts of the world) so many are ill prepared for participating in all that richness, and computing profs tend to stick with teaching in computer science programs so the rest of the world has to figure out things they could help with for themselves.

I think it is about time that we relegated computer science to a minor (not unimportant) stream and got back into the real world – the one with women in it. There’s still a pressing need to bring more women into that minor stream: we need inspirations like Mary Tsingou, we could do worse than preferentially hiring more non-male professors, and we desperately need to shift the discriminatory culture surrounding (especially) mathematics but, if we can at least teach in a way that better represents the richness and diversity of the computing profession itself, it would be a good start.

Originally posted at: https://landing.athabascau.ca/bookmarks/view/10624709/some-thoughts-for-ada-lovelace-day

Educational technology: what it is and how it works | AI & Society

https://rdcu.be/ch1tl

This is a link to my latest paper in the journal AI & Society. You can read it in a web browser from there, but it is not directly downloadable. A preprint of the submitted version (some small differences and uncorrected errors here and there, notably in citations) can be downloaded from https://auspace.athabascau.ca/handle/2149/3653. The published version should be downloadable for free by Researchgate members.

This is a long paper (about 10,000 words), that summarizes some of the central elements of the theoretical model of learning, teaching and technology developed in my recently submitted book (still awaiting review) and that gives a few examples of its application. For instance, it explains:

  • why, on average researchers find no significant difference between learning with and without tech.
  • why learning styles theories are a) inherently unprovable, b) not important even if they were, and c) a really bad idea in any case.
  • why bad teaching sometimes works (and, conversely, why good teaching sometimes fails)
  • why replication studies cannot be done for most educational interventions (and, for the small subset that are susceptible to reductive study, all you can prove is that your technology works as intended, not whether it does anything useful).

Abstract

This theoretical paper elucidates the nature of educational technology and, in the process, sheds light on a number of phenomena in educational systems, from the no-significant-difference phenomenon to the singular lack of replication in studies of educational technologies.  Its central thesis is that we are not just users of technologies but coparticipants in them. Our participant roles may range from pressing power switches to designing digital learning systems to performing calculations in our heads. Some technologies may demand our participation only in order to enact fixed, predesigned orchestrations correctly. Other technologies leave gaps that we can or must fill with novel orchestrations, that we may perform more or less well. Most are a mix of the two, and the mix varies according to context, participant, and use. This participative orchestration is highly distributed: in educational systems, coparticipants include the learner, the teacher, and many others, from textbook authors to LMS programmers, as well as the tools and methods they use and create.  From this perspective,  all learners and teachers are educational technologists. The technologies of education are seen to be deeply, fundamentally, and irreducibly human, complex, situated and social in their constitution, their form, and their purpose, and as ungeneralizable in their effects as the choice of paintbrush is to the production of great art.

Originally posted at: https://landing.athabascau.ca/bookmarks/view/8692242/my-latest-paper-educational-technology-what-it-is-and-how-it-works

Turns out the STEM ‘gender gap’ isn’t a gap at all

Grace Hopper and Univac, image from en.wikipedia.org/wiki/Grace_HopperAt least in Ontario, it seems that there are about as many women as men taking STEM programs at undergraduate level. This represents a smaller percentage of women taking STEM subjects overall because there are way more women entering university in the first place. A more interesting reading of this, therefore, is not that we have a problem attracting women to science, technology, engineering, and mathematics, but that we have a problem attracting men to the humanities, social sciences, and the liberal arts. As the article puts it:

“it’s not that women aren’t interested in STEM; it’s that men aren’t interested in poetry—or languages or philosophy or art or all the other non-STEM subjects.”

That’s a serious problem.

As someone with qualifications in both (incredibly broad) areas, and interests in many sub-areas of each,  I find the arbitrary separation between them to be ludicrous, leading to no end of idiocy at both extremes, and little opportunity for cross-fertilization in the middle. It bothers me greatly that technology subjects like computing or architecture should be bundled with sciences like biology or physics, but not with social sciences or arts, which are way more relevant and appropriate to the activities of most computer professionals. In fact, it bothers me that we feel the need to separate out large fields like this at all. Everyone plays lip service to cross-disciplinary work but, when we try to take that seriously and cross the big boundaries, there is so much polarization between the science and arts communities that they usually don’t even understand one another, let alone work in harmony. We don’t just need more men in the liberal arts – we need more scientists, engineers, and technologists to cross those boundaries, whatever their gender. And, vice versa, we need more liberal artists (that sounds odd, but I have no better term) and social scientists in the sciences and, especially, in technology.

But it’s also a problem of category errors in the other direction. This clumping together of the whole of STEM conceals the fact that in some subjects – computing, say – there actually is a massive gender imbalance (including in Ontario), no matter how you mess with the statistics. This is what happens when you try to use averages to talk about specifics: it conceals far more than it reveals.

I wish I knew how to change that imbalance in my own designated field of computing, an area that I deliberately chose precisely because it cuts across almost every other field and did not limit me to doing one kind of thing. I do arts, science, social science, humanities, and more, thanks to working with machines that cross virtually every boundary.

I suspect that fixing the problem has little to do with marketing our programs better, nor with any such surface efforts that focus on the symptoms rather than the cause. A better solution is to accept and to celebrate the fact that the field of computing is much broader and vastly more interesting than the tiny subset of it that can be described as computer science, and to build up from there. It’s especially annoying that the problem exists at Athabasca where a wise decision was made long ago not to offer a computer science program. We have computing and information systems programs, but not any programs in computer science. Unfortunately, thanks to a combination of lazy media and computing profs (suffering from science envy) that promulgate the nonsense, even good friends of mine that should know better sometimes describe me as a computer scientist (I am emphatically not), and even some of our own staff think of what we do as computer science. To change that perception means not just a change in nomenclature, but a change in how and what we, at least in Athabasca, teach. For example, we might mindfully adopt an approach that contextualizes computing around projects and applications, rather than its theory and mechanics. We might design a program that doesn’t just lump together a bunch of disconnected courses and call it a minor but that, in each course (if courses are even needed), actively crosses boundaries – to see how code relates to poetry, how art can inform and be informed by software, how understanding how people behave can be used in designing better systems, how learning is changed by the tools we create, and so on.

We don’t need disciplines any more, especially not in a technology field. We need connections. We don’t need to change our image. We need to change our reality. I’m finding that to be quite a difficult challenge right now.

 

Address of the bookmark: http://windsorstar.com/opinion/william-watson-turns-out-the-stem-gender-gap-isnt-a-gap-at-all/wcm/ee4217ec-be76-4b72-b056-38a7981348f2

Originally posted at: https://landing.athabascau.ca/bookmarks/view/2929581/turns-out-the-stem-%E2%80%98gender-gap%E2%80%99-isn%E2%80%99t-a-gap-at-all