Anatomies of Intelligence

December 2019, Workshop at NN Cluster initiative hosted by artist-researcher Winnie Soon, curator-researcher Magda Tyzlik-Carver and researcher Pablo Velasco at Aarhus University.

Researchers

Joana Chicau [PT/NL]is a graphic designer, coder, researcher — with a background in dance. Her trans-disciplinary project interweaves web programming languages and environments with choreography. In her practice she researches the intersection of the body with the constructed, designed, programmed environment, aiming at in widening the ways in which digital sciences is presented and made accessible to the public. She has been actively participating and organizing events with performances involving multi-location collaborative coding, algorithmic improvisation, open discussions on gender equality and activism.

joanachicau.com

Jonathan Reus [US/NL] is a musician and artist who explores expanded forms of music-making and improvisational performance through technological artefacts. His practice is cross-disciplinary and research-based, involving open and iterative processes of collaboration with practitioners from across the arts, sciences and humanities. His work tries to confront and challenge the representational capacities of mathematical-logistical systems, algorithms, and infrastructure through a practice of invasive intuition and trust in the diversity of lived experiences.

jonathanreus.com





* Agenda *



  • Welcome to the Anatomies of Intelligence project research: concepts; tools and methodology;
  • Warm-up exercise: Body Atlas, an anatomic journey;
  • Introduction to: data classification and clustering;
  • In focus: K-means clustering and notions of (Euclidean) space and distance;
  • Group exercise The anatomies of a K-means clustering algorithm;
  • Wrapping up: Presentations and Feedback ;





  • Anatomies of Intelligence

    Anatomies of Intelligence is an artistic research initiative seeking to make connections between the formats and collections of anatomical knowledge and investigations into the “anatomy” of computational learning and prediction processes, datasets and machine learning models.

    Continue reading about project research: concepts; tools and methodology.

    Machine Learning

    Machine learning (ML) is the scientific study of algorithms and statistical models that computer systems use to perform a specific task without using explicit instructions, relying on patterns and inference instead. It is seen as a subset of artificial intelligence. [1]

    Supervised learning is the machine learning task of learning a function that maps an input to an output based on example input-output pairs. It infers a function from labeled training data consisting of a set of training examples. In supervised learning, each example is a pair consisting of an input object (typically a vector) and a desired output value (also called the supervisory signal). A supervised learning algorithm analyzes the training data and produces an inferred function, which can be used for mapping new examples. An optimal scenario will allow for the algorithm to correctly determine the class labels for unseen instances. This requires the learning algorithm to generalize from the training data to unseen situations in a "reasonable" way. [1]

    Unsupervised learning is a type of learning that helps find previously unknown patterns in data set. In contrast to supervised learning, unsupervised learning can be used to develop knowledge about the structure and internal relationships of a dataset without pre-existing labels, classifications or categories. Two of the main methods used in unsupervised learning are principal component and cluster analysis. [1]

    Semi-supervised learning (bootstrapping) is a class of machine learning tasks and techniques that also make use of unlabeled data for training – typically a small amount of labeled data with a large amount of unlabeled data. Semi-supervised learning falls between unsupervised learning (without any labeled training data) and supervised learning (with completely labeled training data). Semi-supervised learning is also of theoretical interest in machine learning and as a model for human learning.

    Reinforcement learning is an area of machine learning concerned with how software agents ought to take actions in an "environment" in order to maximize some notion of cumulative reward. It differs from supervised learning in not needing labelled input/output pairs be presented, and in not needing sub-optimal actions to be explicitly corrected. Instead the focus is on finding a balance between exploration (of uncharted territory) and exploitation (of current knowledge).[1]

    Deep learning is part of a broad family of machine learning methods used to learn complex and non-linear models of data, usually through the implementation of various "deep" or heirarchical structures such as a neural network. The learning techniques can be supervised, semi-supervised or unsupervised.

    Data classification and Clustering

    Classification is the problem of identifying to which of a set of categories a new observation belongs, on the basis of a training set of data containing observations (or instances) whose category membership is known. Examples are assigning a given email to the "spam" or "non-spam" class, and assigning a diagnosis to a given patient based on observed characteristics of the patient (sex, blood pressure, presence or absence of certain symptoms, etc.). Classification is an example of pattern recognition. [2]

    Cluster analysis (clustering) is a task in unsupervised learning of grouping a set of objects in such a way that objects in the same group (called a cluster) are more "similar" to each other than to those in other clusters.

    Instead of "learning" from feedback provided by ideal behavior or reinforcement phenomena, cluster analysis identifies commonalities in the data and "learns" in response to the presence or absence of such commonalities in each new piece of data. This approach helps detect anomalous data points that do not fit easily into groups. As the results of clustering, data in the same group has higher similarity metric to each other than to those in other groups, and each group have a centroid to represent the group as a whole. [1]

    K-means and notions of (Euclidean) space & distance


    K-means clustering intends to partition n objects into k clusters in which each object belongs to the cluster with the nearest mean. This method produces exactly k different clusters of greatest possible distinction. The best number of clusters k leading to the greatest separation (distance) is not known as a priori and must be computed from the data. The objective of K-Means clustering is to minimize total intra-cluster variance, or, the squared error function: [3]

    k-means error formula

    is a method of vector quantization, originally from signal processing, that is popular for cluster analysis in data mining. k-means clustering aims to partition n observations into k clusters in which each observation belongs to the cluster with the nearest mean, serving as a prototype of the cluster. This results in a partitioning of the data space into Voronoi cells. k-Means minimizes within-cluster variances (squared Euclidean distances), but not regular Euclidean distances, which would be the more difficult Weber problem: the mean optimizes squared errors, whereas only the geometric median minimizes Euclidean distances. Better Euclidean solutions can for example be found using k-medians and k-medoids.



    image k-means clustering animation


    The K-means Algorithm


      Begin with a dataset, each item (datapoint x) of the dataset should be reduced to a set of N features. To make things easier, we choose to represent each feature by a number from 0.0 and 1.0 (normalization).

      For example, if we choose to describe our data using five features (N=5), one data entry could be described by the set of feature values [0.2, 0.44, 0.87, 0.1, 1.0]

    1. Choose the number of desired clusters K
    2. Choose starting cluster centroids c1-ck: either randomly or by choosing from existing data points.
    3. ITERATE the following until the algorithm convergences on a specific set of clusters, or until the end of a fixed number of ITERATIONS
        ITERATION: For each data point xi:
      1. Calculate the distance from the datapoint to each of the centroids: c1, c2 .. ck);
      2. ASSIGN the datapoint to the cluster represented by the closest centroid
        ITERATION: For each cluster i with centroid ci
      1. Calculate a new centroid for cluster i which is the means (average) of all the datapoints assigned to that cluster.
    4. End.

    K-Means is relatively an efficient method. However, we need to specify the number of clusters, in advance and the final results are sensitive to initialization and often terminates at a local optimum. Unfortunately there is no global theoretical method to find the optimal number of clusters. A practical approach is to compare the outcomes of multiple runs with different k and choose the best one based on a predefined criterion. In general, a large k probably decreases the error but increases the risk of overfitting.

    Dimensionality reduction or dimension reduction, is the process of reducing the number of random variables under consideration by obtaining a set of principal variables. It can be divided into feature selection and feature extraction. [1]

    Distances in mathematics, a metric or distance function is a function that defines a distance between each pair of elements of a set. A set with a metric is called a metric space. [1] A metric induces a topology on a set, but not all topologies can be generated by a metric. A topological space whose topology can be described by a metric is called metrizable. [4]

    "In machine learning, many supervised and unsupervised algorithms use Distance Metrics to understand patterns in the input data. Also, it is used to recognize similarities among the data." [5]

    The Euclidean distance or Euclidean metric is the "ordinary" straight-line distance between two points in N-dimensional Euclidean space. [6]


    Euclidean distance in R2











    Commonly used distance metrics in machine learning are:

  • Euclidean Distance
  • Manhattan distance
  • Minkowski Distance
  • Mahalanobis Distance
  • Cosine Distance
  • Jaccard distance
  • Manhattan distance on a grid

    Figure illustrating Manhattan versus Euclidean distance. The red, blue, and yellow lines all have the same Manhattan length (12), whereas the green (Euclidean) distance between the points has length of about 8.48.



    Catalog API

    // calculate distance between two points
    distance([0.0, 0.0], [6.0, 6.0])

    Simple Catalog Clustering:




    cluster(FEATURES_ARRAY, K, ITERATIONS, USE_BREAKPOINTS)

    catalog.allTags.sort() // show feature names in alphabetical order

    cluster(["knowledge", "materiality", "measurement"], 4, 10, false)

    project(Projections.RADIUS)

    project(Projections.SIMPLE2D)

    cluster(["cutting", "measurement", "spectacle"], 5, 10, true)

    TEMPO = 2

    TEMPO = 0.01

    project(Projections.RADIUS)

    project(Projections.SIMPLE2D)

    also see help()


    Test the clustering using the AoI Theatre JavaScript web api as an observational tool. Open the web console and type the script, as in the example:




    References:


    [1] Glossary of artificial intelligence
    [2] Classification
    [3] K-means clustering algorithm
    [4] Distance metrics (mathematics)
    [5] How to measure distances in machine learning
    [6] Euclidean distance [7] K-means Clustering


    Further references:

    — > History of artificial intelligence
    — > Timeline of machine learning
    — > Article on k-means clustering
    — > Wikipedia article on distances
    — > Machine Learning Modelling and distance metrics

    The anatomies of a K-means clustering algorithm


    Human anatomy planes, 2014 Body measurements, back view, Chinese woodcut, 1443

    [left] Anatomical planes, 2014: including median (red), parasagittal (yellow), frontal or coronal plane (blue) and transverse or axial plane (green). Made with the default MakeHuman model (intermediate gender proportions) with breast size reduced to be more androgynous, posed in standard anatomical position. Blender source available on GitHub.

    [middle] From the book, "The Human Body and Health Revised" by Alvin Davison, published in 1908 by Alvin Davison. This book has a copyright of 1908 by Alvin Davison and one in 1924 by American Book Company. The first copyright was pre-1923 and the copyright was never renewed; therefore, the book should be in the public domain Flickr data on 2011-08-17: Tags: vintage, book, medical, image, illustration, drawing, public domain, copyright free License: CC BY 2.0 User: perpetualplum Sue Clark

    [right] Body measurements, back view, Chinese woodcut, 1443: Woodcut from Tong ren shu xue zhen jiu tu jing (Illustrated Manual of Acupoints on the Bronze Man) by Wang Weiyi, published in 1443 CE (8th year of the Zhengtong reign period of the Ming Dynasty), illustrating the system of proportionate body measurements. This illustration consists of a simple outline drawing of a human body, viewed from the back. Captions on the image indicate the distances between various landmarks on the surface of the body, as an aid in establishing the location of acupoints and the paths of the channels at the back of the body. The units of measurement are the chi and cun, the length of the cun being based on the proportions of the individual's body, i.e. 1 cun = the distance from the base of the middle finger to the end of the crease of the middle joint. 10 cun = 1 chi. See 'Lettering' for further details.

    Exercise: rethinking distances & proximities

    AoI screenshot of theatre interface and clustering

    In groups, briefly analize your dataset; then follow the four main tasks bellow:


    — > (1) Define distance

    Collect and discuss diferent notions around defining 'distance' (or the concept of difference);

    Consider existing methods, but also scales and orientations for observation, affective response & affinity.

    Finally choose a choose a measurement system.


  • What measurements and distances could be used?
  • Is your notion of difference multi-dimensional? What are its dimensions?
  • How do embodied experience inform new thinking around measuring and sensing distances?


  • — > (2) Tagging and Weighting

    Create a collection of tags to be assigned to the data samples/ each entry of the data set.

    Assign the tags and give values to each of them.


  • What vocabularies emerge from the categorization and analysis of the dataset?
  • What can be 'put into words' and what is left unspoken?
  • How does the process influence the knowledge of the current dataset?


  • — > (3) Clustering

    AoI screenshot of theatre interface and clustering

    Execute the k-means clustering algorithm and document each iteration of results;

    — > (4) System representation

    Discuss how to represent the knowledge system you are building. And how to best communicate the process from tags (processual happening on the data collection) to the clustering iterations.

    Consider dimensionality reduction and its implications (what is preserved and what is left out).

    Discuss and list potential entry points for performativity, critical and aesthetic interventions in this algorithmic process.


    Further questions:

  • How to visualize the clustering?
  • What different viewpoints and perpectives and can used?
  • What other ways of making the process sensible can be explored? Which other senses can take part beyond visual? For eg. sonification.


  • New Iteration

    — > (0) Restart

    Experiment the process with another dataset; with different measurement systems; parameters; and analyse how it impacts the clustering.

  • What thoughts emerge after the experiments?
  • What different knowledge ontologies emerge through other geometries and directionalities based in bodily experiences and physicalities?
  • What is a unit of anatomical space and how can that inform embodied approaches to machine learning?




  • Links and resources:

    The Catalogue
    The Theatre

    Code Repository
    Project's classification system: a growing glossary (/expansion of “all tags”)

    Key concepts of anatomical relationship
    History of measurement




    Thank you all for taking part, and looking forward to hearing your feedback!