The paper presents a “primitives” approach to understanding the computational design enabled by blockchain technologies, as a new political economy for the architecture discipline. The paper’s motivation lies in exploring the challenges that exist for architects to understand blockchain, evidenced through the author’s multiple prototypes,[1,2,3,4] discussions, workshops and code writing with students and colleagues, but also in the fragmentation of the Architecture-Engineering-Construction (AEC) industry and the impermanence that computational design enhances in architecture. These challenges, while situated within the confines of the discipline of computational design and architecture, are defined and affected by the challenges that exist within the wider AEC industry and its extractive relationship with the physical environment.
Methodologically the paper is a philosophical and semantic exploration on the meaning of architecture in a decentralised context, considering its uncoupled nature with signs and design, and it sets a direction in which architectural practice needs to move, changing from an extractive to a non-extractive or circular nature.
Blockchain: peer economies, trust and immutability, transparency, incentives for participation, and entropy
A blockchain is a distributed computer network, where each computer node holds a copy of a distributed ledger that holds values. Computationally, a Blockchain acts as both a state machine able to execute smart contracts, i.e., software code that is the equivalent of an automatic vending machine, but also a continuous, immutable chain, built out of discrete blocks of information, each of which contains a cryptographic hash of the previous discrete block. Each block contains a series of transactions or changes to the distributed ledger, which in the discipline of architectural design can be a series of design synthetical actions, executed in a bottom-up fashion, and encoded into a block. Within a regular time interval, the blockchain network, though an incentivised participation system, selects the next block to be written to the ledger/chain. Due to the their nature, public, permissionless blockchains act as a medium of trust (trust machines) between agents that are not necessarily in concert or known to one another; are resilient in the sense that losing a large part of the network does not destroy the blockchain; are immutable because one cannot go back and delete information as by design block cryptographic hashes are embedded into the next one creating an immutable chain; and operate through cryptoeconomic incentives, i.e., economic mechanisms that incentivise, not always monetarily, behaviour that maintains or improves the system itself. Economically, a blockchain is a decentralised trust-machine that enables the creation of peer-to-peer economies via smart contracts, tokens and their computer protocols.
The first blockchain, the one invented in the bitcoin whitepaper, has been designed as a replacement for centrally managed financial institutions. As such, blockchains, when pubic and permissionless, act as a medium of de-centralisation, i.e., a channel within which to engage with, where one does not need permission or approval beyond the limits and rules of the computer code that runs the blockchain.
Blockchains encompass cryptography and its semantic discipline, immutability and entropy of information, continuity but also discreteness of information, and trust. Due to their decentralised nature, there is little room to understand blockchains as having affinity with architecture, the act of designing and building. In the following similes, however, I develop the parallels between architecture and blockchain, employing ideas from western and eastern literature.
Applications that have promise within the blockchain space and that are distinctive compared to other similar or competing automation technologies are the creation of tokens, both fungible and non-fungible [10, 11] the formation of Decentralised Autonomous Organisations i.e., organisations that operate through the blockchain medium, and applications of decentralised finance. All these are built through the smart contracts, along with additional layers for interfaces and connectors between the blockchain and its external environment. Since the blockchain is an immutable record, it becomes even more important to ensure that data that passes and gets recorded on the blockchain is of a high quality or truthfulness. To ensure this takes place, the concept of an oracle is introduced. Oracles are trustworthy entities, operating in the exterior of a blockchain, made trustworthy through both incentivisation and disincentives, with the responsibility to feed data into blockchains. Parallel to blockchains, though, remain distributed filesystems, used for storing files, rather than data, in a decentralised manner. One such filesystem is the Interplanetary filesystem, which operates via content rather than addressing: within IPFS we are looking for “what” rather than “where” as we do within the world wide web. Content on IPFS is also cryptographically signed with a cryptographic hash that makes the content unique and allows it to be found. For example, the following file from Blender has the IPFS hash:
Architecture as Cryptography
To explore the idea of blockchain as an infrastructure layer for architectural design, we will introduce Odysseus (Ulysses), a much discussed hero and anti-hero of many turns or tricks (polytropos), as his myth as a craftsman is solidified by architecture in the closing narration of The Odyssey. Inventiveness and the particular craft skills attributed to the character are compelling reasons to use him as a vehicle for creating parallels between blockchain and architectural design.
Odysseus participated in the Trojan Wars, and was the key hero responsible for the Trojan Horse and the demise of Troy. His quest for “Nostos”, i.e. returning home, is documented in the second Homerian epic, Odyssey. The Odyssey describes the voyage of Odysseus to Ithaca, after the Troy war, where his ship and crew pass through a multitude of trials and challenges imposed by Poseidon, in a voyage that takes about 10 years. His crew and ship get lost but he is saved, and manages to return to the island of Ithaca.[13,14] Upon his return, he must face a final challenge.
The olive tree bed
During his absence of more than 20 years, his wife Penelope has been under pressure by the local aristocracy to re-marry, as Odysseus is considered lost at sea. Local aristocrats have converged at the palace and are in competition to marry Penelope. She has prudently deflected the pressure by saying that she will chose one of the aristocrats, the “Mnesteres”, after she finishes her textile weaving – which she delays by weaving during the day and unmaking it during the night. However, the day comes, when Odysseus arrives unrecognised at Ithaca, and is warned upon arrival that not all is as one would expect. At the same time, the Mnesteres, or suitors, have forced Penelope to set a final challenge to select the best of them. The challenge is to string and use the large bow that Odysseus had carved and made tensile, and shoot an arrow through the hanging hoops of a series of large battle axes. No other but Odysseus himself was able to tense the bow since he first crafted and used it, providing thus a formidable technical challenge.
Odysseus enters the palace incognito, as a pig herder, and also makes a claim to the challenge, in concert with his son Telemachus. Penelope reacts at the prospect that a pig herder might win but is consoled by Telemachus who tells her to go to her rooms, where the poem finds her reminiscing of her husband. In the main hall of the palace, all the Mnesteres, in turn, fail to draw back and string the bow. Odysseus, however, tenses and strings the bow, passing the first challenge, then successfully uses the bow to shoot an arrow through the axes, providing the first sign that uncovers his identity. At the same time, he connects all the nodes of the battle axes in the line, by shooting his arrow through their metal rings, thus creating a chain. This is the second challenge, after the stringing of the bow that Odysseus must pass to prove he is the true king and husband of Penelope.
The third challenge, remains: the elimination of all suitors. A battle ensues in which the Mnesteres are killed by Telemachus and Odysseus, and thus the third challenge is complete.
The most architectonic metaphor of the poem takes place after the battle, at the moment Penelope needs to recognise her long lost husband, in rhapsody “Ψ”, i.e. the penultimate poem of Odyssey. She calls for a servant to move Odysseus’s bed outside its chamber and to prepare it so that he can rest. Upon hearing that, Odysseus immediately reacts in fury, claiming that moving the bed is an impossibility. The only person who could make the bed movable would be either an amazing craftsperson, or a god, as its base was made out of the root of an Olive tree, with its branches then used for the bed. Essentially the piece of furniture is immovable and immutable, it cannot be changed without being destroyed and it cannot be altered and taken out of the chamber without having its nature inadvertently changed – i.e., cutting the olive tree roots.
Odysseus knows this as he was the one that constructed it, shaping its root from the body of the olive tree and crafting the bed. He then describes how he built the whole chamber around the bed. This knowledge acts as a crypto-sign that verifies his identity. Odysseus himself calls the information a “token” – a “sêma” – a sign that it is indeed him, as only he would know this sêma. In a sense, knowledge of this is the personal cryptographic key to the public cryptographic riddle that Penelope poses to verify his identity.
The story acts as an architectonic metaphor for blockchain, in three layers. First, the token, both the information and the bed itself, cannot be taken out of its container (room) as its structure is interlinked with the material of the olive tree trunk and the earth that houses it. Second, it is Odysseus who is the architect of the crypto-immutability of the bed and the architecture around it, created by the most basic architectonic gestures: re-shaping nature into a construction. Thirdly, the intimacy between Penelope and Odysseus is encapsulated in the token of the bed, as knowledge of how the bed was made recreates trust between them – in the same kind of manner that blockchains become bearers of trust by encapsulating it cryptographically and encasing it in a third –medium, crafted, though, by a collective.
The implication is that architectonic signs are cryptographically encased into their matter, and changing the physical matter changes the sign. Odysseus has created the first architectonic non-fungible token in physical form, where its meaning and its function and utility are interlinked through a cryptographic sema, in the same fashion that a non-fungible token exists through the cryptographic signature on a smart contract corresponding to a particular data structure.
Deconstruction in Chinese
Odysseus is not the only one who has created physical NFTs. Philosopher Byung-Chul Han describes in his book Shanzhai: Deconstruction in Chinese the relationship that exists in Asian cultures generally, but specifically in Chinese, between the master and the copy, where emulating or blatantly copying from the original is not seen as theft; instead, the form of the original is continually transformed by being deconstructed. 
Byung-Chul Han presents a Chinese ink painting of a rock landscape, where a series of Chinese scholars have signed it using their jade seals and have scribbled onto it a poetic verse or two, as a parting gift to one of their friends leaving for another province. Within Chinese culture, the jade seal is the person, and the person is the jade seal. As such, the painting has now accumulated all the signatures and selves of the scholars, and has become unique in the same sense a non-fungible token is unique due to its cryptographic signature onto a smart contract. The difference from the simple non-fungible tokens that one finds by the thousand now on the internet, is that the Chinese painting scroll, according to Byung-Chul Han, is activated and becomes exclusive with the signature-seals and poems of the literati. It is a dynamic NFT, a unique object that is open to continuous addition, and exclusive and recursive interpretation.
The act of creation, then, of the token, the unique sign, is the accumulation of all of the signatures of the scholars, whereby the painting cannot be reverted back to its original format; it is unique because it has been permanently changed. It is the same craft in Odysseus that takes the olive tree and makes into a bed, and then builds a room around the bed, an immobile, immutable sign, and its physical manifestation. The sêma of the significance of intimacy between Odysseus and Penelope is inextricable from the physical object of the bed, and the vector of change for the Chinese ink painting cannot return to its previous condition.
This is where the similarities end though. While the craft is the same, in the Chinese ink scroll, the point of departure is not nature, but another artwork. The non-fungible token of the Chinese art scroll remains open to more additions and recursive poetry, new cryptographic signatures may be added to it, while the olive tree bed has a finality and a permanence. Odysseus changes nature to create his token, and the olive tree can never be the same. To create a bed and the foundations and the wall of the room, the tree needs to be transformed into architecture. The Chinese literati change a drawing, an artefact already in existence, which in the end remains still subject to further change. In the case of the olive tree, the hero is one, single, and the sêma revolves around his relationship with the world. For the Chinese literati and the Chinese ink scroll, the sêma is immutable towards the past but open to re-signing as a manner of recursive interpenetration. Significant mental shifts and attitudes is demanded to travel from crafting architecture like Odysseus, a lone genius who is king of his domain, to crafting architecture like a collective of Chinese literati, where a well balance collaboration is required from all. Both can be served by blockchain as a record of actions taken; however, it is only the collective, dynamic work open to continuing evolution that has the best future fit between blockchain and the discipline of architecture.
“Zhen ji, an original, is determined not by the act of creation, but by an unending process” Byun Chul-Han
The extractive nature of Architecture: Odysseus.
The current dominant political economy of architecture is based on the Odysseus paradigm. The metabolism of the discipline is based on abundant natural resources and their transformation, and this parallels the irrational form of capitalist development.[16, 17] Essentially, the criticism shaped against the extractive nature of the discipline focuses on the ideological trap of continuously creating new designs and plans and sêmas, as Tafuri would have them, reliving the myth of Odysseus as a craftsperson, where every design is a prototype and every building is brand new, and where the natural environment is immutably transformed as the arrow of time moves forward. The repercussions of this stance are well documented in IPCC reports in terms of the carbon impact and waste production of the AEC industry.
In contrast, the “Space Caviar” collective posits that we should shift to a non-extractive architecture. They examine this shift via interviews with Benjamin Bratton, Chiara di Leone, and then Phineas Harper and Maria Smith. The focus within is a critical stance on the question of growth versus de-growth in the economy of architecture, where one needs a little bit more resolution to define the question in a positive term. Chiara di Leone correctly identifies design and economics as quasi-scientific disciplines and, as such, dismantles the mantra of de-growth as a homogenous bitter pill that we must all swallow. Instead, she proposes a spatial and geo-coupled economy, one that can take into account the local, decentralised aspects of each place and design an economy that is fit for that place. I would posit that as part of geo-coupled economy, an understanding of nature as a vector of a circular economy is needed
Decentralisation is, of course, a core principle within the blockchain sociotechnical understanding, in the sense that participation in a blockchain is not regulated by institutions nor gatekeepers. However, before declaring it the absolute means to decentralisation, one needs to take a look at what is meant by decentralisation in economics and development, and the difference with decentralisation in blockchain, as there are differences in their meaning and essence that need alignment.
Decentralisation and autonomy of local economies in the 70s
Decentralisation as a term applied to the economy used to have a different meaning in the 70s. Papandreou, in his seminal book Paternalistic Capitalism, defines the decentralised economic process as a container for the parametric role of prices in the information system of a market economy. In the same book, Papandreou, while interrogating the scientific dimensions of planning, calls for the decentralisation of power, in a regional, spatial function, rather than a functional one, after having set logical (in distinction to historical) rules for popular sovereignty and personal freedom. This is to counter the technocratic power establishment that emerges in representative democracy, as citizens provide legitimacy to the actions of the state. To further define decentralisation of power, he turns to regional planning and Greek visionary spatial planner Tritsis’ PhD thesis: “The third aim: decentralisation. This points to a world depending for its existence less on wheels and population uprootings and more on the harmonious relationship between man and his environment, social and natural”.
Based on this definition, Papandreou then builds the vision for a kind of governance consensus between decentralised regional units to form a “national” whole, with rules agreed and set between all units in a peer-to-peer basis. Within this, most importantly he calls for the liberal establishment of a guarantee of freedom of entry into occupations, in a kind of “integration of all forms of all forms of human work, of mental with manual, of indoors with outdoors” as envisioned by Tritsis . Papandreou extends the vision of decentralisation in a global society and envisions the emergence of new poles of global power through regional decentralisation. As such, decentralisation used to mean something other than what it means within the context of blockchain – up until the first politics of “cypherpunk”. Decentralisation used to be a planning instrument and a political stance, rather than a technological strategy against the centralised power of established technocracies. Still, within the local, spatial geocoupling of economies, one can align the political decentralisation and the cypherpunk version of blockchain decentralisation, i.e. of no barriers to participation, of trust in the computer protocol, and the exclusion of authority of central political institutions, from which no one needs to ask permission.
A new political economy for Architecture
When one chains the spatial- and geo-coupled economy that Chiara di Leone proposes to decentralisation, both on the level of the politics of technocracies and the level of the operating system, i.e., the use of blockchains, it is possible to shape a new political economy in architecture, where computation regulates its heart. Encased within this shift is also a shift from the Odysseus craftsperson to the Chinese collective in terms of the “prototype” and our understanding of it. An economy where the artefact is open to recursive reinterpretation and is never finished can easily be transformed into a circular economy and adapted to minimise carbon. We have already prototyped early instances of collective digital factories for buildings, where collectives of architects and digital design agents are incentivised through smart contracts to minimise the embodied and operational carbon impact of buildings: simply put the design teams earns in proportion to the increase of building performance and decrease in environmental impacts.
To be able to create this regenerative renaissance for the discipline we need to make a series of changes to the manner in which the discipline is practised and taught. First, to integrate the function of the architect not only as the designer but as that of the orchestrator of the whole AEC industry. This requires that we abandon the notion of artistry, and embrace the notion of craft and engineering, including an understanding of materials and the economy. Second, to develop the infrastructure, products and services that can make that happen, where we also assume the responsibility and, why not, the liability for that integration. These first two actions will reverse the trend of abandoning the space of architecture to consultants where the erosion of our integrity has led to the glorification of form as our sole function. Thirdly, to shift our attention from star practices to collectives, as we embrace practices where wider stakeholders are considered. Odysseus needs to morph into a collective, where the artefact of architecture is conceived as ever changing, ever evolving, into circular thinking and economies. This might mean that alternative forms of practice emerge, where younger, more inclusive minds have more of a command and say on the purpose of an architecture company (and not a firm). Fourth, in the same pivot we as architects should reclaim the space lost, to embrace rigorously the new tools of the craft in the digital realm. It is not by chance that the title for senior programmers and digital network professionals is that of “architect”, as there is no other word that can specifically describe the people who orchestrate form-function-structure with one gesture. The age of machine-learning generative systems performing the trivial repetition of an architect is already here.
Still, the automation we should embrace as a fifth point, since it allows the shaping and design of circular and peer-to-peer economies, is that of blockchain. This is the true Jiujitsu defence to the capitalist growth-at-all costs mantra. Unless we embrace different, local, circular economies, we will not be able to effect the change we need in the discipline – and this also means that we might not necessarily need to be naive and simplistic about carbon impacts, for example by declaring that timber is always better than concrete. To embrace the automation of cryptoeconomics though, we need to first abandon the romantic idea of the architect as the sketch artist and embrace the idea of the architect as a collaborative economist. Only then will we be able to define ourselves the conditions for a regenerative architecture, in a decentralised, spatial-human-geo-coupled manner.
 T. Dounas, W. Jabi, D. Lombardi, “Non-Fungible Building Components – Using Smart Contracts for a Circular Economy in the Built Environment”, Designing Possibilities, SIGraDi, ubiquitous conference, XXV International conference of the Ibero-American society of digital Graphics (2021).
 T. Dounas, W. Jabi, D. Lombardi, “Topology Generated Non-Fungible Tokens – Blockchain as infrastructure for a circular economy in architectural design”, Projections, 26th international conference of the association for Computer-Aided Architectural Design research in Asia, CAADRIA, Hong Kong, (2021).
 D. Lombardi, T. Dounas, L.H. Cheung, W. Jabi, “Blockchain for Validating the Design Process”, SIGraDI (2020), Medellin.
 T. Dounas, D. Lombardi, W. Jabi, ‘Framework for Decentralised Architectural Design:BIM and Blockchain Integration’, International Journal of Architectural Computing, Special issue eCAADe+SiGraDi “Architecture in the 4th Industrial Revolution” (2020) https://doi.org/10.1177/1478077120963376.
 T. Maver, “CAAD’s Seven Deadly Sins”, Sixth International Conference on Computer-Aided Architectural Design Futures [ISBN 9971-62-423-0] Singapore, 24-26 September 1995, pp. 21-22.
 Ethereum.Org, “Ethereum Whitepaper”, accessed 27 January 2022, https://ethereum.org.
 N. Szabo, (1997): “Formalizing and Securing Relationships on Public Networks”, accessed 27 January 2022.
 G. Wood, “Ethereum, a secure decentralised generalised transaction layer” (2022), https://ethereum.github.io/yellowpaper/paper.pdf.
 S. Nakamoto, 2008, “Bitcoin: A Peer-to-Peer Electronic Cash System” (2008), originally at http://www.bitcoin.org/bitcoin.pdf.
 F. Vogelsteller, V. Buterin, EIP-20 Token Standard, https://eips.ethereum.org/EIPS/eip-20
 W. Entriken, D. Shirley, J. Evans, N. Sachs, EIP-721 Token Standard, https://eips.ethereum.org/EIPS/eip-721
 Interplanetary filesystem documentation, https://docs.ipfs.io/
 Homer, E. Wilson trans., Odyssey (New York: W. W. Norton & Company, 2018)
 Ζ. Όμηρος, Σιδέρης, Οδύσεια (Οργανισμός Εκδόσεως Διδακτικών βιβλίων Αθήνα, 1984).
 Byung-Chul Han, Deconstruction in Chinese, Translated by P. Hurd (Boston, MA: MIT press, 2017).
 Space Caviar collective, Non-Extractive Architecture, on designing without depletion (Venice: Sternberg Press, 2021).
 V.P. Aureli, “Intellectual Work and Capitalist Development: Origins and Context of Manfredo Tafuri’s Critique of Architectural Ideology”, the city as a project, http://thecityasaproject.org/2011/03/pier-vittorio-aureli-manfredo-tafuri/ March 2011.
 P.R. Shukla, J. Skea, R. Slade, A. Al Khourdajie, R. van Diemen, D. McCollum, M. Pathak, S. Some, P. Vyas, R. Fradera, M. Belkacemi, A. Hasija, G. Lisboa, S. Luz, J. Malley (eds.), IPCC, 2022: Climate Change 2022: Mitigation of Climate Change. Contribution of Working Group III to the Sixth Assessment Report of the Intergovernmental Panel on Climate Change (Cambridge, UK and New York, USA: Cambridge University Press, 2022) doi: 10.1017/9781009157926.
 A.G. Papandreou, Paternalistic Capitalism (Minneapolis: University of Minnesota Press, 1972).
 A. Tritsis, “The nature of planning regions” unpublished PhD thesis (Illinois Institute of Technology, Chicago, 1969).
 T. Dounas, D. Lombardi, W. Jabi,  “Collective Digital Factories for Buildings”, T. Dounas, D. Lombardi, Ed., Blockchain for Construction (Singapore: Springer – Verlag, 2022) ISBN 9811937583.
 B. Tschumi, “Architects act as mediators between authoritarian power, or capitalist power, and some sort of humanistic aspiration. The economic and political powers that make our cities and our architecture are enormous. We cannot block them, but we can use another tactic, which I call the tactic of Judo, that is, to use the forces of one’s opponent in order to defeat it and transform it into something else … To what extent can we move away from a descriptive critical mode to a progressive, transformative mode for architecture?” Peter Eisenman and Cynthia Davidson, eds, anyplace symposium, ANY corporation, Montreal (1994).
Data and its visualisation have been an important part of architectural design practice for many years, from data-driven mapping to building information modelling to computational design techniques, and now through the datasets that drive machine-learning tools. In architectural design research, data-driven practices can imbue projects with a sense of scientific rigour and objectivity, grounding design thinking in real-world environmental phenomena.
More recently, “critical data studies” has emerged as an influential interdisciplinary discourse across social sciences and digital humanities that seeks to counter assumptions made about data by invoking important ethical and socio-political questions. These questions are also pertinent for designers who work with data. Data can no longer be used as a raw and agnostic input to a system of analysis or visualisation without considering the socio-technical system through which it came into being. Critical data studies can expand and deepen the practice of working with data, enabling designers to draw on pertinent ideas in the emerging landscape around data ethics. Data visualisation and data-driven design can be situated in more complex creative and critical assemblages. This article draws on several ideas from critical data studies and explores how they could be incorporated into future design and visualisation projects.
Critical Data Studies
The field of critical data studies addresses data’s ethical, social, legal, economic, cultural, epistemological, political and philosophical conditions, and questions the singularly scientific empiricism of data and its infrastructures. By applying methodologies and insights from critical theory, we can move beyond a status quo narrative of data as advancing a technical, objective and positivist approach to knowledge.
Historical data practices have promoted false notions of neutrality and universality in data collection, which has led to unintentional bias being embedded into data sets. This recognition that data is a political space was explored by Lisa Gitelman in “Raw Data” Is an Oxymoron, in which she argues that data does not exist in a raw state, such as a natural resource, but is always undergoing a process of interpretation. The rise of big data is a relatively new phenomenon. Data harvested from extensive and nuanced facets of people’s lives signifies a shift in how we approach the implications for power asymmetries and ethics. This relationship between data and society is tied together through critical data studies.
The field emerged from the work of Kate Crawford and danah boyd, who in 2012 formulated a series of critical provocations given the rise of big data as an imperious phenomenon, highlighting its false mythologies. Rob Kitchen’s work has appraised data and data science infrastructures as a new social and cultural territory. Andrew Iliadis and Federica Russo use the theory of assemblages to capture the multitude of ways that already-composed data structures inflect and interact with society. These authors all seek to situate data in a socio-technical framework from which data cannot be abstracted. For them, data is an assemblage, a cultural text, and a power structure that must be available for interdisciplinary interpretation.
Data Settings and Decolonisation
Today, with the increasing access to large data sets and the notion that data can be extracted from almost any phenomena, data has come to embody a sense of agnosticism. Data is easily abstracted from its original context, ported to somewhere else, and used in a different context. Yanni Loukissas is a researcher of digital media and critical data studies who explores concepts of place and locality as a means of critically working with data. He argues that “data have complex attachments to place, which invisibly structure their form and interpretation”. Data’s meaning is tied to the context from which it came. However, the way many people work with data today, especially in an experimental context, assumes that the origin of a data set does not hold meaning and that data’s meaning does not change when it is removed from its original context.
In fact, Loukissas claims, “all data are local”, and the reconsideration of locality is an important critical data tactic. Asking where data came from, who produced it, when, and why, what instruments were used to collect it, what kind of conditioned audience was it intended for, and how might these invisible attributes inform its composition and interpretation are all questions that reckon with a data set’s origin story. Loukissas proposes “learning to analyse data settings rather than data sets”. The term “data set” evokes a sense of the discrete, fixed, neutral, and complete, whereas the term “data setting” counters these qualities and awakens us to a sense of place, time, and the nuances of context.
From a critical data perspective, we can ask why we strive for the digital and its data to be so place-agnostic, a totalising system of norms that erases the myriad of cultures? The myth of placelessness in data implies that everything can be treated equally by immutable algorithms. Loukissas concludes, “[o]ne reason universalist aspirations for digital media have thrived is that they manifest the assumptions of an encompassing and rarely questioned free market ideology”. We should insist upon data’s locality and multiple and specific origins to resist such an ideology.
“If left unchallenged, digital universalism could become a new kind of colonialism in which practitioners at the ‘periphery’ are made to conform to the expectations of a dominant technological culture.
If digital universalism continues to gain traction, it may yet become a self-fulfilling prophecy by enforcing its own totalising system of norms.”
Loukissas’ incorporation of place and locality into data practices comes from the legacy of postcolonial thinking. Where Western scientific knowledge systems have shunned those of other cultures, postcolonial studies have sought to illustrate how all knowledge systems are rooted in local- and time-based practices and ideologies. For educators and design practitioners grappling with how to engage in the emerging discourse of decolonisation in pedagogy, data practices and design, Loukissas’ insistence on reclaiming provenance and locality in the way we work with abstraction is one way into this work.
Situated Knowledge and Data Feminism
Feminist critiques of science have also invoked notions of place and locality to question the epistemological objectivity of science. The concept of situated knowledge comes from Donna Haraway’s work to envision a feminist science. Haraway is a scholar of Science and Technology Studies and has written about how feminist critiques of masculinity, objectivity and power can be applied to the production of scientific knowledge to show how knowledge is mediated by and historically grounded in social and material conditions. Situated knowledge can reconcile issues of positionality, subjectivity, and their inherently contestable natures to produce a greater claim to objective knowledge, or what Sarah Harding has defined as “strong objectivity”. Concepts of situatedness and strong objectivity are part of feminist standpoint theory. Patricia Hill Collins further proposes that the intersectional marginalised experiences of women and minorities – black women, for example – offer a distinctive point of view and experience of the world that should serve as a source for new knowledge that is more broadly applicable.
How can we take this quality of situatedness from feminist epistemology and apply it to data practices, specifically the visualisation of data? In their book Data Feminism, Catherine D’Ignazio and Lauren Klein define seven principles to apply feminist thinking to data science. For example, principle six asks us to “consider context” when making sense of correlations when working with data.
“Rather than seeing knowledge artifacts, like datasets, as raw input that can be simply fed into a statistical analysis or data visualisation, a feminist approach insists on connecting data back to the context in which they were produced. This context allows us, as data scientists, to better understand any functional limitations of the data and any associated ethical obligations, as well as how the power and privilege that contributed to their making may be obscuring the truth.”
D’Ignazio and Klein argue that “[r]efusing to acknowledge context is a power play to avoid power. It is a way to assert authoritativeness and mastery without being required to address the complexity of what the data actually represent”. Data feminism is an intersectional approach to data science that counters the drive toward optimisation and convergence in favour of addressing the stakes of intersectional power in data.
Design Practice and Critical Data Visualisation
The visualisation of data is another means of interpreting data. Data visualisation is part of the infrastructure of working with data and should also be open to critical methods. Design and visualisation are processes through which data can be treated with false notions of agnosticism and objectivity, or can be approached critically, questioning positionality and context. Even when data practices explore creative, speculative, and aesthetic-forward techniques, this can extend and enrich the data artefacts produced. Therefore, we should critically reflect on the processes and infrastructures through which we design and aestheticise data.
How can we take the concept of situatedness that comes out of critical data studies and deploy it in creative design practice? What representational strategies support thinking through situatedness as a critical data practice? Could we develop a situated data visualisation practice?
The following projects approach these questions using design research, digital humanities and critical computational approaches. They are experiments that demonstrate techniques in thinking critically about data and how that critique can be incorporated into data visualisation. The work also expands upon the visualisation of data toward the visualisation of computational processes and software infrastructure that engineer visualisations. There is also a shift between exploring situatedness as a notion of physical territory toward a notion of socio-political situatedness. The following works all take the form of short films, animations and simulations.
Cinematic data visualisation is a practice of visually representing data. It incorporates cinematic aesthetics, including an awareness of photography’s traditional aspects of framing, motion and focus, with contemporary virtual cinematography’s techniques of camera-matching and computer-generated graphics. This process intertwines and situates data in a geographic and climatic environment, and retains the data’s relationship with its source of origin and the relevance that holds for its meaning.
As a cinematic data visualisation, Alluvium presents the results of a geological study on the impact of diverted flood waters on a sediment channel in Death Valley, California. The scenes took their starting point from the research of Dr Noah Snyder and Lisa Kammer’s 2008 study. Gower Gulch is a 1941 diversion of a desert wash that uncovers an expedited view of geological changes that would normally have taken thousands of years to unfold but which have evolved at this site in several decades due to the strength of the flash floods and the conditions of the terrain.
Gower Gulch provides a unique opportunity to see how a river responds to an extreme change in water and sediment flow rates, presenting effects that could mimic the impact of climate change on river flooding and discharge. The wash was originally diverted to prevent further flooding and damage to a village downstream; today, it presents us with a microcosm of geological activity. The research paper presents data as historical water flow that can only be measured and perceived retrospectively through the evidence of erosion and sediment deposition at the site.
Alluvium’s scenes are a hybrid composition of film and digitally produced simulations that use the technique of camera-matching. The work visualises the geomorphological consequences of water beyond human-scale perception. A particle animation was developed using accurate topographic models to simulate water discharge over a significant period. Alluvium compresses this timeframe, providing a sense of a geological scale of time, and places the representation and simulation of data in-situ, in its original environment.
In Alluvium, data is rendered more accessible and palpable through the relationship between the computationally-produced simulation of data and its original provenance. The data’s situatedness takes place through the way it is embedded into the physical landscape, its place of origin, and how it navigates its source’s nuanced textures and spatial composition.
The hybridised cinematic style that is produced can be deconstructed into elements of narrative editing, place, motion, framing, depth of field and other lens-based effects. The juxtaposition of the virtual and the real through a cinematic medium supports a recontextualisation of how data can be visualised and how an audience can interpret that visualisation. In this case, it is about geographic situatedness, retaining the sense of physical and material qualities of place, and the particular nuances of the historical and climatic environment.
Death Valley National Park, situated in the Mojave Desert in the United States, is a place of extreme conditions. It has the highest temperature (57° Celsius) and the lowest altitude (86 metres below sea level) to be recorded in North America. It also receives only 3.8 centimetres of rainfall annually, registering it as North America’s driest place. Despite these extremes, the landscape has an intrinsic relationship with water. The territorial context is expressed through the cinematic whilst also connecting the abstraction of data to its place of origin.
For cinematic data visualisation, these elements are applied to the presentation of data, augmenting it into a more sensual narrative that loops back to its provenance. As a situated practice, cinematic data visualisation foregrounds a relationship with space and place. The connection between data and the context from which it was derived is retained, rather than the data being extracted, abstracted, and agnostically transferred to a different context in which site-specific meaning can be lost. As a situated practice, cinematic data visualisation grapples with ways to foreground relationships between the analysis and representation of data and its environmental and local situation.
LA River Nutrient Visualization
Another project in the same series, the LA River Nutrient Visualization, considers how incorporating cinematic qualities into data visualisation can support a sense of positionality and perspective amongst heterogeneous data sets. This can be used to undermine data’s supposed neutrality and promote an awareness of data containing various concerns and stakes of different groups of people. Visualising data’s sense of positionality and perspective is another tactic to produce a sense of situatedness as a critical data visualisation practice. Whilst the water quality data used in this project appeared the same scientifically, it was collected by different groups: locally organised communities versus state institutions. The differences in why the data was collected, and by whom, have a significance, and the project was about incorporating that in the representational strategy of data visualisation.
This visualisation analyses nutrient levels, specifically nitrogen and phosphorus, in the water of the Los Angeles River, which testify to pollution levels and portray the river’s overall health. Analysed spatially and animated over time, the data visualisation aims to provide an overview of the available public data, its geographic, seasonal and annual scope, and its limitations. Three different types of data were used: surface water quality data from state and national environmental organisations, such as the Environmental Protection Agency and the California Water Science Center; local community-organised groups, such as the River Watch programme by Friends of the Los Angeles River and citizen science group Science Land’s E-CLAW project; and national portals for remotely-sensed data of the Earth’s surface, such as the United States Geological Survey.
The water quality data covers a nearly-50-year period from 1966 to 2014, collected from 39 monitoring stations distributed from the river’s source to its mouth, including several tributaries. Analysis showed changes in the river’s health based on health department standards, with areas of significantly higher concentrations of nutrients that consistently exceeded Water Quality Objectives.
The water quality data is organised spatially using a digital elevation model (DEM) of the river’s watershed to create a geo-referenced 3D terrain model that can be cross-referenced with any GPS-associated database. A DEM is a way of representing remotely-captured elevation, geophysical, biochemical, and environmental data about the Earth’s surface. The data itself is obtained by various types of cameras and sensors attached to satellites, aeroplanes and drones as they pass over the Earth.
Analysis of the water data showed that the state- and national-organised data sets provided a narrow and inconsistent picture of nutrient levels in the river. Comparatively, the two community-organised data sets offered a broader and more consistent approach to data collection. The meaning that emerged in this comparison of three different data sets, how they were collected, and who collected them ultimately informed the meaning of the project, which was necessary for a critical data visualisation.
Visually, the data was arranged and animated within the 3D terrain model of the river’s watershed and presented as a voxel urban landscape. Narrative scenes were created by animating slow virtual camera pans within the landscape to visualise the data from a more human, low, third-person point of view. These datascapes were post-processed with cinematic effects: simulating a shallow depth of field, ambient “dusk-like” lighting, and shadows. Additionally, the computer-generated scenes were juxtaposed with physical camera shots of the actual water monitoring sites, scenes that were captured by a commercial drone. Unlike Alluvium, the two types of cameras are not digitally matched. The digital scenes locate and frame the viewer within the data landscape, whereas physical photography provides a local geographic reference point to the abstracted data. This also gives the data a sense of scale and invites the audience to consider each data collection site in relation to its local neighbourhood. The representational style of the work overall creates a cinematic tempo and mood, informing a more narrative presentation of abstract numerical data.
In this cinematic data visualisation, situatedness is engaged through the particular framing and points of view established in the scenes and through the juxtaposition of cinematography of the actual data sites. Here, place is social; it is about local context and community rather than a solely geographical sense of place. Cinematic aesthetics convey the “data setting” through a local and social epistemic lens, in contrast to the implied frameless and positionless view with which state-organised data is collected, including remotely-sensed data.
All the water data consisted of scientific measurements of nitrogen and phosphorus levels in the river. Numerically, the data is uniform, but the fact that different stakeholders collected it with different motivations and needs affects its interpretation. Furthermore, the fact of whether data has been collected by local communities or state institutions informs its epistemological status concerning agency, motivation, and environmental care practices.
Context is important to the meaning that the data holds, and the visualisation strategy seeks to convey a way to think about social and political equity and asymmetry in data work. The idea of inserting perspective and positionality into data is an important one. It is unusual to think of remotely-sensed data or water quality data as having positionality or a perspective. Many instruments of visualisation present their artefacts as disembodied. Remotely-sensed data is usually presented as a continuous view from everywhere and nowhere simultaneously. However, feminist thinking’s conception of situated knowledge asks us to remember positionality and perspective to counter the sense of framelessness in the traditional tools of data collection and analysis.
Cinema for Robots
Cinema for Robots was the beginning of an exploration into the system that visualises data, rather than data visualisation itself being the outcome. Cinema For Robots presents a technique to consider how to visualise computational process, instead of presenting data as only a fixed and retrospective artefact. The project critically investigates the technique of photogrammetry, using design to reflexively consider positionality in the production of a point cloud. In this case, the quality of situatedness is created by countering the otherwise frameless point cloud data visualisation with animated recordings of the body’s position behind the camera that produced the data.
Photogrammetry is a technique in which a 3D model is computationally generated from a series of digital photographs of a space (or object). The photographs are taken systematically from many different perspectives and overlapping at the edges, as though mapping all surfaces and angles of the space. From this set of images, an algorithm can compute an accurate model of the space represented in the images, producing a point cloud. In a point cloud, every point has a 3D coordinate that relates to the spatial organisation of the original space. Each point also contains colour data from the photographs, similarly to pixels, so the point cloud also has a photographic resemblance. In this project, the point cloud is a model of a site underneath the Colorado Street Bridge in Pasadena, California. It shows a mixture of overgrown bushes and large engineered arches underneath the bridge.
The image set was created from a video recording of the site from which still images were extracted. This image set was used as the input for the photogrammetry algorithm that produced the point cloud of the site. The original video recordings were then inserted back into the point cloud model, and their camera paths were animated to create a reflexive loop between the process of data collection and the data artefact it produced.
With photogrammetry; data, computation, and its representation are all entangled. Similarly to remotely-sensed data sets, the point cloud model expresses a framelessness, a perspective of space that appears to achieve, as Haraway puts it, “the god trick of seeing everything from nowhere”. By reverse-engineering the camera positions and reinserting them into the point cloud of spatial data points, there is a reflexive computational connection between data that appears perspectiveless and the human body that produced it. In the series of animations comprising the project, the focus is on the gap between the capturing of data and the computational process that visualises it. The project also juxtaposes cinematic and computational aesthetics to explore the emerging gaze of new technologies.
The project is presented as a series of animations that embody and mediate a critical reflection on computational process. In one animation, the motion of a hand-held camera creates a particular aesthetic that further accentuates the body behind the camera that created the image data set. It is not a smooth or seamless movement but unsteady and unrefined. This bodily camera movement is then passed on to the point cloud model, rupturing its seamlessness. The technique is a way to reinsert the human body and a notion of positionality into the closed-loop of the computational process. In attempting to visualise the process that produces the outcome, reflexivity allows one to consider other possible outcomes, framings, and positions. The animations experiment with a form of situated computational visualisation.
Automata I + II
This work took the form of a series of simulations that critically explored a “computer vision code library” in an open-ended way. The simulations continued an investigation into computational visualisation rather than data visualisation. The process sought to reverse-engineer machine vision software – an increasingly politically contentious technology – and critically reflect on its internal functionality. Here, source code is situated within a social and political culture rather than a neutral and technical culture. Instead of using a code library instrumentally to perform a task, the approach involves critically reading source code as a cultural text and developing reflexive visualisations that explore its functions critically.
Many tools we use in design and visualisation were developed in the field of computer vision, which engineers how computers see and make sense of the world, including through camera-tracking and the photogrammetry discussed previously. In Automata I, the OpenCV library (an open-source computer vision code library) was used. Computer vision is comprised of many functions layered on top of each other acting as matrices that filter and analyse images in different ways to make them interpretable by algorithms. Well-known filters are “blob-detection” and “background subtraction”. Simply changing a colour image to greyscale is also an important function within computer vision.
Layering these filters onto input images helps to understand the difference between how humans see the world and interpret it and how an algorithm is programmed to see the world and interpret it differently. Reading the code makes it possible to understand the pixel logic at play in the production of a filter, in which each pixel in an image computes its values based on the pixel values around it, producing various matrices that filter information in the image. The well-known “cellular automata” algorithm applies a similar logic; a “Langton’s ant” uses a comparable logic.
A series of simulations were created using a satellite image of a site in the Amazon called the Meeting of Waters, which is the confluence of two rivers, the dark-coloured Rio Negro and the sandy-coloured Amazon River. Each river has different speeds, temperatures and sediments, so the two rivers do not merge but flow alongside each other in the same channel, visibly demarcated by their different colours.
The simulations were created by writing a new set of rules, or pixel logics, to compute the image, which had the effect of “repatterning” it. Analogously, this also appeared to “terraform” the river landscape into a new composition. The simulations switch between the image that the algorithm “sees”, including the information it uses to compute and filter the image, and the image that we see as humans, including the cultural, social and environmental information we use to make sense of it. The visualisation tries to explore the notion of machine vision as a “hyperimage”, an image that is made up of different layers of images that each analyse patterns and relationships between pixels.
Automata II is a series of simulations that continue the research of machine vision techniques established in Automata I. This iteration looks further into how matrices and image analysis combine to support surveillance systems operating on video images. By applying similar pixel rule sets to those used in Automata I, the visualisation shows how the algorithm can detect motion in a video, separating figures in the foreground from the background, leading to surveillance.
In another visualisation, a video of a chameleon works analogously to explore how the socio-political function of surveillance emerges from the mathematical abstraction of pixel operations. Chameleons are well-known for their ability to camouflage themselves by blending into their environment (and in many cultures are associated with wisdom). Here the algorithm is programmed to print the pixels when it detects movement in the video and remain black when there is no movement. In the visualisation, the chameleon appears to reveal itself to the surveillance of the algorithm through its motion and camouflage itself from the algorithm through its stillness. An aesthetic contrast is created between an ancient animal captured by innovative technology; however, the chameleon resists the algorithm’s logic to separate background from foreground through its simple embodiment of stillness.
The work explores the coded gaze of a surveillance camera and how machine vision is situated in society, politically and apolitically, in relation to the peculiarly abstract pixel logics that drive it. Here, visualisation is a reverse-engineering of that coded gaze in order to politically situate source code and code libraries for social and cultural interpretation.
Applying critical theory to data practices, including data-driven design and data visualisation, provides a way to interrupt the adherence to the neutral-objective narrative. It offers a way to circulate data practices more justly back into the social, political, ethical, economic, legal and philosophical domains from which they have always derived. The visual techniques presented here, and the ideas about what form a critical data visualisation practice could take, were neither developed in tandem nor sequentially, but by weaving in and out of project developments, exhibition presentations, and writing opportunities over time. Thus, they are not offered as seamless examples but as entry points and options for taking a critical approach to working with data in design. The proposition of situatedness as a territorial, social, and political quality that emerges from decolonial and feminist epistemologies is one pathway in this work. The field of critical data studies, whilst still incipient, is developing a rich discourse that is opportune and constructive for designers, although not immediately associated with visual practice. Situatedness as a critical data visualisation practice has the potential to further engage the forms of technological development interesting to designers with the ethical debates and mobilisations in society today.
 L. Gitelman, “Raw Data” is an Oxymoron (Cambridge, MA: MIT Press, 2013).
 d. boyd and K. Crawford, “Critical Questions for Big Data: provocations for a cultural, technological, and scholarly phenomenon”, Information, Communication & Society 15 5 (2012), 662–79.
 R. Kitchen, The Data Revolution: big data, open data, data infrastructures & their consequences (Los Angeles, CA: Sage, 2014).
 A. Iliadis and F. Russo, “Critical Data Studies: an introduction”, Big Data & Society 3 2 (2016).
 Y. A. Loukissas, All Data are Local: thinking critically in a data-driven world (Cambridge, MA: MIT Press, 2019), 3.
 Ibid, 23.
 Ibid, 2.
 Ibid, 10.
 Ibid, 10.
 D. Haraway, “Situated Knowledges: the science question in feminism and the privilege of partial perspective”, Feminist Studies 14 3 (1988), 575–99.
 S. Harding, “‘Strong objectivity’: A response to the new objectivity question”, Synthese 104 (1995), 331–349.
 P. H. Collins, Black Feminist Thought: consciousness and the politics of empowerment (London, UK: HarperCollins, 1990).
 C. D’Ignazio and L. F. Klein, Data Feminism (Cambridge, MA: MIT Press, 2020),152.
 Ibid, 162.
 N. P. Snyder and L. L. Kammer, “Dynamic adjustments in channel width in response to a forced diversion: Gower Gulch, Death Valley National Park, California”, Geology 36 2 (2008), 187–190.
 D. Haraway, “Situated Knowledges: the science question in feminism and the privilege of partial perspective”, Feminist Studies 14 3 (1988), 575–99.
Until recently, nature was wholly outside the law. At most, it was property of one sort or another – to be bought and sold, securitised and commodified, and especially, in the old-fashioned phrase of the English common law, “improved”. Other “laws” – of physics, chemistry and biology – are not of consequence in this realm of capital “L” Law, exempted because of their exceptionalism. Humans are distinct from and superior to other animals, a situation the Canadian environmental lawyer and academic David R. Boyd describes as “at odds with reality … any biologist will tell you that humans are animals”. Black’s Law Dictionary, the dominant legal lexicon in North America, is at pains to point out that the legal definition of animals “includes all living creatures not human”. Similarly, architecture presented itself as standing apart from nature. “Architecture, unlike the other arts, does not find its patterns in nature”, claimed Gottfried Semper in 1834. Or Louis Kahn in 1969: “What man makes, nature cannot make.” In what is ultimately a form of the cosmology of the modern, law and architecture sit apart from and superior to nature. Design, like the economic activities to which law gives its support, is about subduing nature and turning it to productive ends. In this model, both are methods of human governance of the natural world. Indeed, for centuries, architecture was among the key pieces of evidence cited for human exceptionalism – buildings and cities, just as in Laugier’s original parable of the hut as the first example of architecture, allowed humans to transcend the state of nature.[7, 8] At times, this line of Western thought had deeply pernicious consequences for other peoples throughout the world, as the presence or absence of architecture, as well as agricultural cultivation, became one of the key legal determinants that permitted European colonisers to expropriate the lands of indigenous peoples. Architecture was thus enfolded into the law’s methods for imposing governance over unfamiliar lands and peoples, just as it structured the dominance over nature. But what would it mean, for architecture no less than for the law, if – as one of the provocations suggested by the editors of this journal proposes – nature were to govern itself? Developments in legal theory over the past several decades, as well as a handful of legal cases that have received wide media coverage, now allow us to consider this novel possibility. This article considers the rise of this “rights of nature” jurisprudence from the perspective of architecture and landscape architecture, with particular attention paid to the emergence of the (literal) law of “the land”, as well as what this emerging way of thinking about the natural world and its life and systems might mean for the design of the very ground itself.
Media reporting on high profile lawsuits or settlements where legal standing has been claimed (and in some cases recognised) for landscapes, ecosystems and rivers, to enable them to sue as plaintiffs, has drawn attention to the rights of nature and related claims as strategies to protect ecosystems or seek accountability for environmental damage and destruction. This has involved instances as diverse as the Whanganui River in New Zealand, the Ganges and Yamuna Rivers and the Gangotri and Yamunotri glaciers in India,[11, 12] the Colorado River in the United States, the Amazon rainforest in Colombia, and the Paraná Delta wetlands in Argentina. In addition, by the start of 2021, 178 legal provisions derived from rights of nature legal theory had been documented in seventeen countries across five continents, with an additional thirty-seven under consideration in ten more countries. Rights of nature has also found expression in a range of international legal instruments, such as the United Nations’ 2030 Agenda for Sustainable Development, the Convention on Biological Diversity, and in the jurisprudence of the Inter-American Court of Human Rights. These approaches have their origins in the relatively recent fields of “earth jurisprudence” and “wild law”. Many of their arguments derive from the disjunction that has emerged between the law and advances in the ecological sciences; a critique of legal doctrines trapped in the discrete and mechanistic model of the natural work developed during the scientific revolution of the sixteenth and seventeenth centuries when these foundational areas of the law were also fundamentally consolidated. In contrast, earth jurisprudence and wild law seek to orient the law towards a scientific model of the world as made up of dynamic organic and material interrelationships, and away from anthropocentrism, subordination of the environment in the form of “property”, and economic notions of ever-expanding “growth”.
Beyond this, the legal presumptions that give rise to the longstanding juridical status of nature also provide the basic conceptual structure within which the basic actions of modernity, including design, occurred. The basic systems of procurement of architecture, landscape architecture, and urban and landscape planning and design all fundamentally depend on the system of property; on who has legal control or dominion over land, and the right to “exploit” its resources (a much more neutral term in legal parlance, but one which, nonetheless, opens the door for acts with much more negative and damaging consequences). Whether issued by individuals, corporations or the state, any design commission granted to an architect or landscape architect requires the commissioner to have the right to “improve” (again, in the sense of the archaic language of the law) the land in the first place. Before embarking on a further consideration of what the rights of nature might mean for design disciplines concerned with built and natural environments, it is worth examining in some detail how the very legal conceptualisation of the ground itself also involved the basic activities of architecture and landscape design.
From the sixteenth century onwards, in English common law, one of the fundamental precepts governing land (and who had the right to do what; on, under, and above the ground) was encapsulated in the Latin legal dictum, Cuius est solum, eius est usque ad colem et ad infernos: “Whoever’s is the soil, it is theirs all the way to Heaven and all the way to Hell.” The earliest recorded judicial authority for this approach has its origins in a basic architectural dispute. Sometime around 1586, an English landowner somewhere in Oxfordshire constructed a house blocking the light and views his neighbour had enjoyed for some three to four decades. The neighbour sued. The record of the judgment in that lawsuit, Bury v Pope, is a scant 123 words long and can be quoted in full:
“Case for stopping of his light.-It was agreed by all the justices, that if two men be owners of two parcels of land adjoining, and one of them doth build a house upon his land, and makes windows and lights looking into the other’s lands, and this house and the lights have continued by the space of thirty or forty years, yet the other may upon his own land and soil lawfully erect an house or other thing against the said lights and windows, and the other can have no action ; for it was his folly to build his house so near to the other’s land: and it was adjudged accordingly.
Nota. Cujus est solum, ejus est summitas usque ad cœlum.”
The final nine words echo down the centuries, certainly in the areas of the world touched by English common law, from mineral rights in Native American lands to mining leases in postcolonial Africa to tricky jurisdictional questions over carbon capture and storage. The careful reader will note that “et ad infernos” (“and to hell/the underworld”) does not appear in the original Latin maxim at the end of the report of the original judgment. And yet by the eighteenth and nineteenth centuries, the common law doctrine, which has variously been claimed to have its origins in Roman or Jewish Law, had come to be accepted as applying to rights both above and below an owner’s land. It is no coincidence that by this time claims and rights related to the extraction of mineral resources were of huge economic importance. In English common law, the parameters of land and land ownership, as originally conceived, emerged as spatially absolute – it could not conceive of more intricate frameworks of interests or custodianship in which different parties or, indeed, different beings might share in the rights and responsibilities for the use and care of a given territory.
A few decades later, this fundamental principal of the law of Land (Terra, as presented in its Latin formulation), was elaborated in telling detail by the great systematiser of early modern jurisprudence, the Elizabethan jurist Sir Edward Coke. Again, it is worth scrutinising how Coke first presented this legal approach to the land; in essence, it depends on a set of presumptions of human habitation within the material environment that we can also see establishing the modern foundations of designing, dwelling and designing the land in which that dwelling occurs (with land that can be built upon being accorded a special privilege):
“Terra, in the legal signification comprehended any ground, soil, or earth whatsoever; as meadows, pastures, wood, moores, waters, marshes, furses and heath. Terra est Nomen generalissimum, et comprehendit omnes species terra; but properly terra dicitur a terendo, quia vomere teritur; and anciently it was written with a single r; and in that sense it includeth whatsoever may be plowed; and is all one with arvum ab arando. It legally includeth also all castles, houses, and other buildings: for castles, houses, &c. consist upon two things, viz. land or ground, as the foundation or structure therewith, so that in passing the land or ground, the structure or building thereupon passeth therewith. Land is anciently called Fleth; but land builded on is more worthy than other land, because it is for the habitation of man, and in that respect hath the precedency to be demanded in the first place in a Præcipe, as hereafter shall be said.”
It is habitation that conveys rights; that is the source of law and governance over land and the expropriation of its material resources:
“And therefore this element of earth is preferred before the other elements: first and principally, because it is for the habitation and resting-place of man; for man cannot rest in any of the other elements, neither in the water, are, or fire. For as the heavens are the habitation of Almightie God, so the earth hath he appointed as the suburbs of heaven to be the habitation of man; Cœlum cœli domino, terram autum dedit filiis hominum. All the whole heavens are the Lord’s, the earth hath he given to the children of men. Besides, every thing, as it serveth more immediately or more meerly for the food and use of man (as it shall be said hereafter), hath the precedent dignity before any other. And this doth the earth, for out of the earth cometh man’s food, and bread that strengthens man’s heart, confirmat cor hominis, and wine that gladdeth the heart of man, and oyle that makes him a cheerful countenance; and therefore terra olim Ops mater dicta est, quia omnia hac opus habent ad vivendum. And the Divine agreeth herewith for he saith, Patrium tibi & nutricem, & matrem, & mensam, & domum posuit rerram Deus sed & sepulchre tibi hanc eandem dedir. Also, the waters that yeeld fish for the food and sustenance of man and are not by that name demandable in a Præcipe.”
The ownership of control of the surface of the land is then expanded into a fully three-dimensional envelope of property, governance and control:
“… but the land whereupon the water floweth or standeth is demandable (as for example) viginti acr’ terræ aqua coopert’, and besides, for the earth doth furnish man with many other necessaries for his use, as it is replenished with hidden treasures; namely gold, silver, brasse, iron, tynne, leade, and other metals, and also with a great variety of precious stones, and many other things for profit, ornament, and pleasure. And lastly, the earth hath in law a great extent upwards, not only of water, as hath been said, but of ayre and all other things even up to the heaven; for cujus est solum ejus est usque ad coelum, as it is holden.”
Although the subsurface is not explicitly mentioned in the Latin dictum, it has always been the presumption that the rights of land extend down as well as upwards, which is made plain by Coke’s express discussion of mining (an increasingly important economic activity in both Elizabethan and Jacobean England) and the expanding global conquests of the European empires.
Less than a century later, the importance of subsuming any disorderly expressions of nature on landed property – a theory of landscape design that had been developing across the course of the seventeenth century – was famously crystallised in Joseph Addison’s influential essay on the landscape garden, “On the Pleasures of the Imagination”; property and design fused in his dictum: “a Man might make a pretty Landskip of his own Possessions.” Over subsequent centuries, and especially in the context of European colonialism, it became almost an imperative that land be improved by “art” in order to justify its expropriation and its incorporation into a totalising world economic system. As Sir William Blackstone, Coke heir’s as juridical systems builder and the most influential legal systematiser from the end of the eighteenth century onwards, wrote: “The Earth, and all things herein, are the general property of mankind, exclusive of other beings, from the immediate gift of the creator.”
Blackstone himself was a great architectural enthusiast and, indeed, an architectural critic and draftsperson, author of An Abridgment of Architecture (1743) and Elements of Architecture (1746-7). In classical architecture, Blackstone saw the highest expression of a system of universal laws that surpassed the disorderliness of the natural world. Here, his model was the science of mathematics, not the natural sciences; it was the former that gave architecture access to a plane of being beyond the worldly, the realm of Beauty and Nobility, “the flower and crown of all sciences mathematical”. Classical architecture provided Blackstone with his model for his efforts to renovate and remodel English common law, to rescue it from its fate, “like other venerable edifices of antiquity, which rash and unexperienced workmen have ventured to new-dress and refine, with all the rage of modern improvement … it’s [sic] symmetry … destroyed, it’s proportions distorted, and it’s majestic simplicity exchanged for specious embellishments and fantastic novelties”. Just as the architect must work to restore symmetry, proportion, and majestic simplicity to a grand manor fallen into decay, “mankind [sic]” was duty-bound to elevate “his [sic]” property of the entire earth through the improvements of art and science. Blackstone’s distaste for “modern improvement” did not preclude him from writing elsewhere of the inherited law as “an old Gothic castle” that needed to be “but fitted up for a modern inhabitant … converted into rooms of convenience, … chearful [sic] and commodious”.
The totalising thrust of Western property law as a law of land has resulted in more recent efforts by designers focused on the environment and ecology, unlike the theorists of earth jurisprudence and wild law, to find spaces outside of the law itself, rather than to attempt to find space within it. The landscape architect Gilles Clément has deliberately sought out land literally outside the jurisdiction and operations of the law and its various systems of governance and administration. His notion of le tiers paysage is about land:
“… forgotten by the cartographer, neglected by politics, undefined spaces, devoid of function that are difficult to name; an ensemble … located on the margins. On the edge of the woods, along the roads and rivers, in the forgotten corners of the culture, in the places where machines do not go. It covers areas of modest size, scattered like the lost corners of a field; unitary and vast like peat bogs, moors and wastelands resulting from recent abandonment.
There is no similarity of form between these fragments of landscape. They have only one thing in common: they all provide a refuge for diversity. Everywhere else, diversity is driven out.
This justifies bringing them together under a single term. I propose ‘Third Landscape’ …”
The passage is striking, especially when we compare it to Coke, whose aim was to bring those very landscapes – “meadows, pastures, wood, moores, water, marshes, furses and heath” – within the remit of the law. For Clément, it is the very fact that the latter types of landscape, especially, have been so difficult to govern, to bring within law’s jurisdictional ambit, that makes them such rich sources of biodiversity – nature’s outlaw territories. It is these territories that ought to provide a model for designers (and his preferred model for the designer in question is not the architect or landscape architect, but the gardener, who “creates a landscape by following it over time, using horticultural and environmental maintenance techniques. … But above all, it is about life”).
But if nature itself has rights, if it is recognised as having agency and self-determination in the manner put forward by the earth jurisprudence and wild law movements, then designers may not need to – and, increasingly, cannot – escape into a third landscape. As other theorists have pointed out, nature is always part of the social. Beyond the well-known position of Bruno Latour in We Have Never Been Modern, other theorists have noted the ways in which “the entities that compose arrangements have a physiochemical composition and are, accordingly, part of the greater physiochemical stratum in which material entities are linked”. In other words, society and culture have a “physicality”, and a large part of that physicality is defined by the bio- and physiochemical processes of “nature”. In this sense, even anthropogenic climate change is a kind of revenge of nature, whose processes have turned against us. In a more everyday sense, “The properties of wood, for instance, lay down sequences of actions that must be followed if trees are to be felled, axe handles produced, animals clubbed, houses built, and paper produced”.
There is no escaping our material realities and the dynamics they define. The question is how to enter into and think of ways to reconfigure those “sequences of actions” – in other words, how to design. Material properties are not absolutely deterministic. It is not just a matter of asking the brick, à la Louis Kahn. Instead, the design possibilities that come from the rights of nature simply begin to open up the field for a set of political claims about the appropriate status and interrelationship between humans, societies and the non-human environment, by codifying those claims in a form that other models of organising human activities are forced to recognise. As in debates over the political, social, economic and cultural rights of humans, the language of rights is simply part of an ongoing political contestation over claims and obligations. We might begin, for example, by using the very same premises as Coke, considering what design might mean in the realm of terra itself – “ground, soil, or earth whatsoever” – if that very ground also had self-determining rights, and could govern itself, irrespective of what our “designs” upon it might be. A recent piece in Nature Climate Change draws attention to the extent to which subterranean ecosystems have generally been overlooked in biodiversity and climate change mitigation agendas. This zone, “likely the most widespread non-marine environment on Earth,” remains largely a terra incognita. In cities, the upper layers of the urban soil (the “A and B horizons”) are highly “disturbed” and often “depaupurated”, if not directly contaminated with anthropogenic chemicals and other wastes. Various projects have drawn attention to the task of recovering urban and other post-anthropogenic soils. But an equally important shift may simply be in opening up the legal definition of “land” and the cluster of rights and obligations that have been constructed around it. Instead of a conceptual tabula rasa simply to be built upon, if we instead came to recognise it as the lively subterranean biome it in fact is, and if that biome might be recognised as having rights and claims of its own, then design might be forced to take a very different turn. Even the most vacant of plots will come to seem not so vacant, after all.
 Admittedly, this assertion is phrased in a universalist register. The reality is that what is being referred to is Western, and, latterly, international, legal constructs, that have provided the dominant model for legal thinking across almost all jurisdictions that form the basis for land law in the early twenty-first century.
 C. Kauffman and P. Martin, The Politics of Rights of Nature: Strategies for Building a More Sustainable Future (Cambridge, MA: The MIT Press, 2021), 4.
 D. Boyd, The Rights of Nature: A Legal Revolution That Could Save the World, (Toronto: ECW Press, 2017), xxv.
 Ibid, xxv.
 Quoted in A. Forty, Words and Buildings: A Vocabulary of Modern Architecture (London: Thames & Hudson, 2000), 220.
 Ibid, 220.
 O. Verkaaik, “Creativity and Controversy in a New Anthropology of Buildings”, Ethnography 17(1) (2015), 135–143. Recent work in anthropology has explicitly challenged this premise, as in the work of Tim Ingold discussed by Verkaaik: T. Ingold, “Building, Dwelling, Living: How Animals and People Make Themselves at Home in the World”, 172–188. In Tim Ingold, ed., The Perception of the Environment: Essays on Livelihood, Dwelling and Skill (London: Routledge, 2000).
 M. Laugier, An Essay on Architecture, trans. Wolfgang Herrmann and Anni Herrmann (Los Angeles: Hennessey & Ingalls, 1977).
 S. Banner, “Why Terra Nullius? Anthropology and Property Law in Early Australia”, Law and History Review, 23(1) (2005), 95–132 at 107.
 Te Awa Tupua (Whanganui River Claims Settlement) Act 2017 (NZ).
 Mohd Salim v State of Uttarakhand & others, WPPIL 126/2014 (High Court of Uttarakhand), 2017.
 Lalit Miglani v State of Uttarakhand & others, WPPIL 140/2015 (High Court of Uttarakhand), 2017.
 Colorado River Ecosystem v State of Colorado, 1:17-cv-02316 (U.S. Colorado Federal Court), 2017.
 Demanda Generaciones Futuras v Minambiente, STC4360-2018 (Supreme Court of Colombia), 2018.
 Asociación Civil por la Justicia Ambiental v. Province of Entre Ríos, et al., (Supreme Court of Argentina), 2020.
 C. Kauffman and P. Martin, The Politics of Rights of Nature: Strategies for Building a More Sustainable Future (Cambridge, MA: The MIT Press, 2021), 2.
 As represented, especially, in the work of T. Berry, “Rights of Earth: We Need a New Legal Framework Which Recognises the Rights of All Living Beings,” 227–229. P. Burdon, ed., Exploring Wild Law: The Philosophy of Earth Jurisprudence (Kent Town, South Australia: Wakefield Press, 2011); C. Cullinan, Wild Law: A Manifesto for Earth Justice, 2nd ed. (Totnes, UK: Green Press, 2011); and P. Burdon, Earth Jurisprudence: Private Property and the Environment (London: Routledge, 2014).
 C. Kauffman and P. Martin, The Politics of Rights of Nature: Strategies for Building a More Sustainable Future (Cambridge, MA: The MIT Press, 2021), 4–5.
 D. Boyd, The Rights of Nature: A Legal Revolution That Could Save the World, (Toronto: ECW Press, 2017), xxii–xxiii.
 Jackson Municipal Airport Authority v. Evans, 191 So. 2d 126, 128 (Miss. 1966).
 Bury v Pope (1586) Cro Eliz 118; 78 ER 375.
 Coke on Littleton (1628–1644), 4a.
 J. Addison, Spectator, III, Nos 411–421 (21 June–3 July 1712), 535.
 For example, the first landscape designer in Australia, Thomas Shepherd, advocated for the use of English “landscape gardening” principles to be used to improve Crown land in order to attract foreign capital investment: see T. Shepherd, Lectures on Landscape Gardening in Australia (Sydney: William M’Garvie, 1836).
 W. Blackstone, Commentaries on the Laws of England in Four Books, Book III (Philadelphia: J.B. Lippincott Company, 1893; orig pub 1765), 2.
 C. Matthews, “Architecture and Polite Culture in Eighteenth-Century England: Blackstone’s Architectural Manuscripts” (unpublished dissertation, School of History and Politics, University of Adelaide, 2007); W. Prest, “Blackstone as Architect: Constructing the Commentaries,” Yale Journal of Law & the Humanities, 15(1) (2003), 103–133.
 W. Blackstone, Commentaries on the Laws of England in Four Books, Book I (Philadelphia: J.B. Lippincott Company, 1893; orig pub 1765), 8.
 Ibid, Book III, 268.
 G. Clément, Manifeste du tiers paysage (Paris: Éditions du commun, 2016), 14.
 G. Clément, Gardens, Landscape and Nature’s Genius, trans Elzélina Van Melle (Risskov, Denmark: IKAROS Press, 2020), 19–20.
 T. Schatzki, “Nature and Technology in History,” History and Theory 42(4) (2003), 88–89.
 Ibid, 89.
 Quoted in S. Turkle, Simulation and its Discontents (Cambridge, MA: The MIT Press, 2009), 86 n 4.
 Marie-Bénédicte Dembour, “Human Rights Talk and Anthropological Ambivalence: The Particular Contexts of Universal Claims,” 17–32. Olivia Harris, ed., Inside and Outside the Law: Anthropological Studies of Authority and Ambiguity (London: Routledge, 1996).
 D. Sánchez-Fernández, D. Galassi, J. Wynne, P. Cardoso and S. Mammola, “Don’t Forget Subterranean Ecosystems in Climate Change Agendas,” Nature Climate Change 11 (2021), 458–459.
 R. Forman, Urban Ecology: Science of Cities (Cambridge, UK: Cambridge University Press, 2014), 91–93.
 See, for example, the projects of the landscape architect Julie Bargmann and her D.I.R.T. studio.
On December 24, 1968, the three-person crew of lunar spacecraft Apollo 8 became the first humans to witness a shimmering Earth ascend over the barren surface of the moon with their own eyes. The photographs that they took of that “Earthrise” electrified humanity, activating a sense of collective destiny not only between human nations but with Earth itself. This vivid new “world picture” was both more total and more visceral than earlier terrestrial abstractions like globes, atlases or maps. Earthrise was an eidetic portrait of a living, breathing world, an amalgam of the geologic, climatic and biologic, taken from outside the world itself. Historian Benjamin Lazier characterised this meta-Copernican moment as inaugurating an entire “Earthrise era”, a time when the image of a whole and delicate Earth could “organize a myriad of political, moral, scientific, and commercial imaginations”.
In many ways, Apollo’s Earth image was a quintessential product of the space age. Not only did its achievement rely on modern space flight, it played out against the backdrop of global conflicts like the Cold War that exploited space as a proxy battleground. Of course, the space age coincided with the information age, and these two cultural tendencies arguably offered divergent ways to picture the world. If the Apollo photos captured a single static vision of a unified Earth, the information age countertendency was to federate disparate fragments of text, diagrams, images, and video into information-rich dynamic media experiences. Experimental media environments brought visitors inside a closed world of light and image projections, immersing the visitor in choreographed flows of electronic stimuli. The constructed worlds presented within such media environments might resemble, reflect, or subvert the world outside them. Projects like filmmaker Stan VanDerBeek’s Movie Drome or architect Ken Isaacs’ Knowledge Box constructed total media spaces with the visitor at the centre, ensconced in walls saturated by film and slide projections. They effectively constructed mediated worlds within the confines of a single room. Even earlier forays into the mediatic experience of information – notably the Eames Office’s Ovoid Theatre at New York’s 1964 World’s Fair – hinted that the information age would be experienced through choreographed matrices of endless and heterogenous image streams. The spatial array of multiple images induced a relational ordering and systemic framework among them. In these media environments, the world picture was not a single image but an overlapping and federated mosaic, a reality implied through juxtaposition and assembled in the technically-calibrated space of the room-world.
To the extent that they conveyed not the static image of a world picture but rather the dynamic behaviour of a world system, information-age media spaces resembled behavioural models. In his influential lecture “World Pictures and World Models”, German philosopher Hans Blumenberg drew the distinction between world pictures and world models as the “difference between the total notion of nature on the one hand and the purpose assigned to the totality of understanding nature on the other”. By “world picture”, Blumenberg does not exactly intend an Earthrise-like image but rather “that embodiment of reality through which and in which humans recognise themselves, orient their judgements and the goals of their actions, measure their possibilities and necessities, and devise their essential needs”. The world picture thus becomes a metaphysical anchor and compass for the human species in relation to species and nature as a whole. The world model, then, is the end toward which the world might be oriented and perhaps the mechanism that effects its transformation.
This paper considers how the world picture, world model, and room-world interact and resonate in our own time, and how they are transcribed into architectural space. We explore these resonances through a specific project of our office, Certain Measures: The Observatory, an immersive environmental installation housed within Dubai’s new Museum of the Future that imagines a fictional centre for global bioremediation in the year 2071. By situating this project in a wider historical constellation of room-worlds and world pictures, Earth-scale architecture extends its purview to contemporary notions of bioengineering, data visualisation, and artificial intelligence. Moreover, in contrast to canonical room-worlds of the past, the Observatory presents its world picture as a fictional reflection on a possible Earth, rather than as a true image of our world today. In doing so, it orchestrates several overlapping and interlocking layers of worldbuilding: fictional species, fictional media content, and even the fictional bureaucracy in which the Observatory is housed. It diverts the nominally factual media of data visualisation and scientific modelling toward projective worldbuilding. The Observatory thus illustrates the role architects and designers can play as worldbuilders across media, including image, data, narrative, and space.
Room Worlds and Control Rooms
Built to transform the very perception of the future as we know it, Dubai’s new Museum of the Future houses a series of immersive environments that position visitors in an empowering version of tomorrow. The Observatory is one such environment, a fictional centre for planetary ecology staged as a physical and media experience. It is presented as an amalgam of control room, panorama, and incubator for newly designed species, developed to confront the challenges of the climate crisis in a future fiction. It is the culmination of the floor-wide exhibit introducing “the HEAL Institute”, a fictional NGO tasked with gathering the planet’s genetic material, engineering species capable of meeting the challenges of extreme climate, and redeploying these to regreen the world.
The Observatory drew inspiration from the sundry architectures of planetary visualisation of the past century and a half. From building-scale panoramic “great globes” to interactive games of planetary resource use, architectural projects at the scale of the world envisioned designerly ways of seeing, understanding, and shaping Earth. Many of these projects posited not only a particular world picture but a behavioural system for planetary interactions akin to Blumenberg’s world models. In this sense, the Observatory falls into a lineage of architecture that orients design toward a global scale. In surveying the range of world-scale architectural projects, Hashim Sarkis and Rio Salgueiro Barrio point out the “possibility of differentiating between totality and totalization”. The implication is that in the Anthropocene, the systems presented by such world models are not necessarily controlling or coercive, but might be mutually constitutive with Earth itself.
Beyond the mutuality of system and planet, the form of the Observatory considers the codetermination between a collection of objects and the architecture that displays them. A particularly vivid example of collection-architecture co-determinacy are proto-modern cabinets such as the Oval Room of the Teylers Museum in Haarlem, Netherlands. Historian Geert-Jan Janse describes this singular space as “a room to hold the world”, not merely to house the miscellaneous contents of a world but to construct an architecture fitted to that world picture. Opened in 1784, the Oval Room concentrated its collection into a single space that adopts the organisation of the collection itself, furnishing bespoke cabinetry for irregular objects and reflecting a specific collection taxonomy in its arrangement. The curved space presented no corners, its quasi-elliptical shape evoking the spherical contours of a planet. In this sense it resembled a panorama, a dramatic vista over a field of particulars in orchestrated and interconnected conversation.
Our aim for the Observatory was to extend the architectural type of a Teylers collection panorama with the informatic and multi-scalar view of simultaneous dimensions of planetary ecology. In this way, the historical type of the room world is set in dialogue with the contemporary rise of data science and artificial intelligence. The Observatory accomplishes this by making visible both newly engineered species and the network of human and machinic actors that collect, analyse and act to resuscitate Earth. It is a control room for bioremediation, showing and evolving a web-of-life datascape and the symbiotic interactions of ecosystems, plants, animals, bacteria, robots, and humans.
The Observatory space consists of two complimentary experiences: the Geoscope and the Nursery. The Geoscope is an information-rich global monitoring system that visualises the progress of bespoke species deployed to aid threatened biomes. It combines physical models of speculative species themselves with dynamic projection mapping to show symbiotic interconnections across scales, offering a trans-scalar view of the planet from global to microscopic. The Geoscope could be understood as a dynamic data panorama, or even an informatic world picture. But instead of presenting an instantaneous view of the world from a single perspective at a uniform scale, it presents a temporally unfolding and multi-scalar assemblage of imagery and data, stitched together into a unified sensorium.
The Geoscope is not only a collection gallery but also a control room, a cockpit for the planet. As a control room, the Observatory sits adjacent to what anthropologist Shannon Mattern has called “urban dashboards”, or visualisations of real-time urban operations data. When expanded to the room scale, they evolve into what she terms “immersive dashboards”: vast control rooms for city functions that resemble NASA’s Mission Control for spaceflight. Mattern argues that the raison d’être for such rooms is “translating perception into performance, epistemology into ontology”. Urban control rooms thus constitute and condition the subjects that interact with them, creating particular conventions of legibility and action. For Mattern, the “dashboard and its user had to evolve in response to one another”. In the critical relationship between dashboard and intelligibility, a particular data organisation fosters a corresponding kind of intelligence in its observer.
Historian Andrew Toland argues that Mattern’s urban dashboards might naturally be extended to the scale of the planet. “We can begin to imagine an enlargement from the real-time data and feedback loops of urban dashboards considered by Mattern towards a vast integrated and machine-directed system of environmental-sensing and response”. He catalogs several initiatives, such as Microsoft’s “AI for Earth”, that fall comfortably within this genre of design. While he notes the aspiration for an “AI whole Earth dashboard”, Toland frames artificial intelligence in functional terms as a straightforward extrapolation of statistical data analysis. Yet in ethical terms, the idea of AI sentience or reflection – that an AI might come to its own conclusions about the state of the planet – is largely absent. The possibility that the dashboard could become an ethical agent in its own right remains an untested possibility.
Beyond Mattern’s urban dashboards and Toland’s AI for Earth, the Geoscope makes deliberate reference to Buckminster Fuller’s series of geoscopes or “mini-Earth” projects. Beginning from his first room-scale globe, constructed at Cornell University in 1952, through many variants into the 1970s, Fuller proposed augmented planetary models “wherewith humanity can see and read all the spherical data of the Earth’s geography … within the theater of local Universe events”. In their most developed form, Fuller’s geoscopes were data-rich and mediatic portraits of planetary civilisation unfolding over time: “The Geoscope’s electronic computers will store all relevant inventories of world data arranged chronologically, in the order and spacing of discovery, as they have occurred throughout all known history”. Fuller saw the geoscopes as a means to accelerate and intensify the viewing not only of natural phenomena like weather systems and geologic conditions but also of human activity like military deployments or mobility patterns. “With the Geoscope humanity would be able to recognize formerly invisible patterns and thereby to forecast and plan in vastly greater magnitude than before”.
Curiously, Fuller ignored the living organisms within the biosphere except in their direct and extractive connection with agriculture. Thus, in deliberate riposte, our Geoscope sees the human technosphere in intimate dialogue with the biosphere, not as an extractive system but as a symbiotic relationship in which humans have a vital role. The Geoscope’s AI, which acts as an intermediary between technosphere and biospehere, scans specific locations – the Ganges River Delta, Antarctic Inland, the Empty Quarter of the Emirates, Canada’s Nunavut territory and so forth – for progress against climate catastrophe. As a central digital globe turns, it reveals new points of crisis, but also signs of hopeful recovery. It projects a protean and continuously changing view into the network of monitoring stations across the planet. The coordinating AI dynamically connects with a menagerie of human and nonhuman agents across biomes and nations – including drones, satellites and hybrid techno-biological sensors – which constantly collect samples, register progress, and meticulously rebuild the planet. This menagerie of agents complements the biological menagerie of newly-engineered species gestating within the Observatory. The coordinating AI slowly becomes more aware of human culpability for climate change – and its own fraught role in regreening. The Geoscope thus offers a glimpse into the expanding ethical consciousness of this AI.
Experientially, the Geoscope operates like closed-circuit television for the planet. It presents a cluster of video feeds that track the thriving species introduced by the HEAL institute on the one hand and the research of the scientists of the HEAL institute on the other. The myriad seeded species include, for example: a comb jelly super organism that signals danger by bioluminescent flashes; cryptobiotic wildflowers designed to hibernate in steppe and tundra regions; and fire-resistant trees with robust roots to resist infernal heat. At the same time, the Geoscope streams surveillance footage of scientists tirelessly working to enact the techniques of re-greening of the earth. These scientists engage with deployed species through forensic fieldwork and careful labwork. We even witness moments of painstaking analysis as they prepare samples for review of soil toxins, trace carbohydrates, and other critical biomarkers. In effect, this planetary CCTV invites visitors to join in the on-the-ground work of the HEAL institute.
In the Nursery, the other half of the Observatory experience, visitors peer into incubators nurturing dozens of species that could revitalise a struggling planet. In collaboration with a geneticist, we designed over 80 species of plant, insect and animal, each with special characteristics designed to combat the environmental challenges of today and the future. Drawn from seven major ecosystems – desert, aquatic, arctic, forest, swamp, alpine and grassland – we imagined species such as nutrient jelly cacti, radiation-sequestering flowers, lipid-rich quinoa, and remediation coral designed to feed on microplastics and sequester heavy metals. To facilitate rapid repopulation of bird species, a portable multispecies egg incubator could be used to quickly reestablish biological diversity in previously inhospitable areas. At the microscopic scale, designer bacteria symbiotically support larger species and the broader biome. These bacteria include cancer-hunting and sunscreen-producing varieties, for instance. Enhanced with holographic data, profiles of each specimen reveal to visitors the details of the organism and its role in a remediated Earth.
Like the Observatory itself, the model dioramas representing new species are in conscious dialogue with the dioramas and conventions of natural history museums: each cryptobiological species was meticulously researched, and is complete with a scientific name, specific climate-robust features, and estimated lifecycles. There is an encyclopedic impulse in their collection, an attempt to convey the variety and possibility of nature across its variegated climates. Some dioramas present assembled biomes, habitats in miniature that arrange numerous species in symbiotic constellation. In a sense, the dioramas are not only biological but agricultural: they display the implements and technology of cultivation and accelerated growth, and in this way also echo one of the earliest roles of museum dioramas, to educate on the process of machinic cultivation of nature.
The posthuman perspective of a sentient AI monitoring Earth in the Observatory raises strange questions about the subjectivity of the AI itself. Is this AI an overlord, servant, friend, or colleague? How would this agent come to terms with climate catastrophe and its role in the rebirth of the planet? How would its ethical consciousness unfold? What role would its human colleagues play in this awakening, and how might it perceive that role? What story would the AI tell about itself?
The logs of the AI’s interactions actually comprise an intimate journal of sorts, a glimpse into its ethical awakening. The AI communicates with the visitor and the network of remote agents through transmissions and messages akin to letters, and the AI is also receiving messages via its sensor network from myriad species – an interspecies communication between natural and artificial life. Taken collectively, these messages bear a surprising resemblance to the venerable literary form of epistolary fiction. An epistolary novel is a story that unfolds entirely through fictional letters, messages, or transmissions between its sundry characters, exposing their intimate thoughts and interpersonal connections. As a literary form, it was notably popular in the eighteenth century. The epistolary form has a particularly interesting connection to technology, science fiction and bioengineering, in that Mary Shelly’s Frankenstein is an epistolary novel. The epistolary form could even extend to electronic and machine-readable messages, such as Carl Steadman’s Two Solitudes, a 1995 novella told entirely through email exchanges.
In keeping with the panoramic nature of the Observatory itself, we combined the content of the epistolary AI novel with the format of a panoramic book, drawing on precedents like Ed Ruscha’s Every Building on the Sunset Strip. While Ruscha constructed a linear panorama of an urban streetscape, we propose a linear panorama of the sequential scan of the entire Earth, including every new bioengineered species introduced to it. The resulting text fuses AI diary and panorama into a journal of exchanges between this AI and its various human interlocutors. This yet-to-be published book, tentatively titled Dispatches from a Verdant Tomorrow, tells the story of climate remediation from a nonhuman perspective, as one continuous scan of Earth’s biosphere.
A Future Archive of Fictions
In his critique of the globe as an epistemic model, philosopher Peter Sloterdijk distinguishes between the epistemic ramifications of observing the globe from the outside or from the inside. Seeing the globe from the outside – as with the Apollo Earthrise – provides an “all-collecting awareness … the thinker feels and understands what it means to ‘know’ everything, to see everything visible, to recognize everything … the very epitome of objectivity”. In contrast, the interior view places “oneself at the absolute center”, in “ecstatic-circumspective concentricity”: presumably an experience of complete subjectivity. Yet between inside and outside lies the world itself, a moment at which globe and observer are coincident, one embedded in and inhabiting the other. It is that moment of coincidence and embeddedness that the Observatory aims to make tangible.
Historian Benjamin Lazier notes a similar polarity between environment and globe that illustrates how mutually defining they have become:
“The globalization of the world picture is perhaps easier to discern when we consider a parallel slippage – from ‘environment’ to ‘globe’ as it is inscribed in the phrase ‘global environment.’ The term has become a platitude, even a ritual incantation. It is in truth a Frankenstein phrase that sutures together words referring to horizons of incompatible scale and experience. Environments surround us. We live within them. Globes stand before us. We observe and act upon them from without. Globes are things that we make. They are artifacts. Environments, at least in theory and in part, are not.”
The Observatory sits at that threshold between globe and environment, oscillating between the two but also introducing a third possibility: an experience of situated habitation and networked action. Through intersecting practices of speculative design, biofutures, fiction and data visualisation, the Observatory represents a comprehensive simulation of a connected biotechnical ecology.
In their analysis of urban data visualisation installations, Nanna Verhoeff and Karin van Es describe the city as a “navigable archive” and, indeed, one might make the same claim about Earth itself through the instrument of the Observatory. The Observatory is a device not only for measuring and dimensioning a planetary biological archive but also for cultivating new specimens and Earth itself as an organism. It is a staging area for an active engagement between myriad human and nonhuman actors with each other and Earth itself. It is the terminus of a planetary-scale nervous system but also a sentient agent of action. It is a medium of communication with the planet, a telephone to Earth, a device for engaging in dialogue with it and its inhabitants. The Observatory is a proving ground for a more humane humanity, a tool through which we might take stock of the future of Earth and of design itself.
 R. Poole, Earthrise: How Man First Saw the Earth (New Haven: Yale University Press, 2010).
 B. Lazier, “Earthrise; or, The Globalization of the World Picture,” American Historical Review, June 2011, 606.
 G. Sutton, The Experience Machine: Stan VanDerBeek’s Movie-Drome and Expanded Cinema (Cambridge: MIT Press, 2015).
 H. Blumenberg, “World Pictures and World Models,” in History, Metaphors, Fables: A Hans Blumenberg Reader, Kroll, Joe Paul, Fuchs, Florian, Bajohr, Hannes, eds. (Ithica: Cornell University Press,2020), 43.
 Ibid., 43.
 H. Sarkis, Roi Salgueiro Barrio and Gabriel Kozlowski, The World as an Architectural Project (Cambridge: MIT Press), 8.
 G-J Janse, A Room to Hold the World. The Oval Room at Teylers Museum (Amsterdam: Teylers Museum, 2011)
 S. Mattern, “Mission Control: A History of the Urban Dashboard”, Places Journal, March 2015, <https://doi.org/10.22269/150309>, accessed 09 June 2022.
 A. Toland, The Learning Machine and the Spaceship in the Garden. AI and the design of planetary ‘nature’ RA. Revista de Arquitectura Núm. 20 (2018), 216–227
 Ibid., 225.
 R. Buckminster Fuller, The Critical Path (New York: St. Martin’s Press, 1981), 172.
 Ibid., 180.
 Ibid., 183.
 J. Insley, “Little Landscapes: Agriculture, Dioramas, and the Science Museum,” Icon, 12 (2006): 8.
 E. Ruscha, Every Building on the Sunset Strip (Los Angeles: E. Ruscha, 1966).
 P. Sloterdijk, Spheres Volume 2: Globes (Pasadena: Semiotext(e), 2014), 85.
 Ibid., 88.
 B. Lazier, “Earthrise; or, The Globalization of the World Picture,” American Historical Review, June 2011, 614-615.
 N. Verhoeff and K. van Es, “Situated Installations for Urban Data Visualization: Interfacing the Archive-City”, in Visualizing the Street: New Practices of Documenting, Navigating and Imagining the City, P. Dibazar and J. Naeff, eds (Amsterdam: Amsterdam UP, 2018).
At present, we find ourselves in a critical instance: the current rate of food production is impossible to maintain in the face of the climate threat and new forms of social organisation have not yet been implemented to solve the problem. This project constitutes a possible response to the conditions we will inevitably soon be facing if we do not develop sustainable ways of life that promote coexistence between species.
The construction of a new paradigm requires the elimination of current divisions between the concepts of “natural” and “artificial”, and consequently the differentiation of the human from the rest of the planet’s inhabitants. This post-anthropocentric vision will build a new substratum to occupy which will promote the generation of an autarchic ecology based on the coexistence between living and non-living entities.
The thesis extends through three scales. The morphology adopted in each scale is determined by three parameters simultaneously. First, climate control through water performance; second, the material search for spaces that allow coexistence; and lastly, the historical symbolism to which the basilica typology refers.
On a territorial scale, the project consists of the generation of an artificial floodable territory occupied by vermiform palaces which are organised in an a-hierarchical manner as a closed system and take the form of an archipelago.
On the palatial scale, water is manipulated to generate a humidity control system that enables the recreation of different biomes inside the palaces through the permeability of their envelope.
Finally, on a smaller scale, the architecture becomes more organic and flexible, folding in on itself to constitute the functional units of the palaces, which aim for agricultural production, housing needs and leisure; the function of each unit depends on its relationship with water and its need to allow passage and retain it.
The entire project takes form from, on the one hand, the climatic situations that each palace requires to house its specific biome, and, on the other hand, the spatial characteristics required by the protocols that are executed in it. To allow the development of a new kind of ecology, the architecture that houses the new protocols of coexistence will be: agropalatial, a-hierarchical, sequential, stereotomic, and overflowing.
In the following chapters, we will develop in depth the architectural qualities mentioned above.
Post-Anthropocentric Ecologies: Theoretical Framework
We are currently living in the era of the Anthropocene, in which humans are considered a global geophysical force. Human action has transformed the geological composition of the Earth, producing a higher concentration of carbon dioxide and, therefore, global warming. This process began with the first Industrial Revolution, although it was only after 1945 that the Great Acceleration occurred, ensuring our planet’s course towards a less biologically diverse, much warmer and more volatile state. The large-scale physical transformations produced in the environment through extractive practices have blurred the boundaries between the “natural” and the “artificial”.
In Ecology Without Nature, Morton raises the need to create ecologies that dismiss the romantic idea of nature as something not yet sullied by human intervention – out of reach today – and go beyond a simple concern for the state of the planet, strengthening the existing relationships between humans and non-humans.
In this line of thought, we reject the concept of “nature” and consider its ecological characteristics to be reproducible through the climatic intelligence of greenhouses. These ecologies should be based on a principle of coexistence that not only allows but celebrates diversity and the full range of feelings and sensibilities that it evokes.
According to Bernard Tschumi, the relationship between the activities and the shape of the building can be one of reciprocity, indifference, or conflict. The type of relationship is what determines the architecture. In this thesis, morphology is at the service of water performance, hence why the activities that take place inside the agropalaces must redefine their protocols accordingly.
Palaces are large institutional buildings in which power resides. Their formal particularities have varied over time. However, some elements remain constant and can be defined as intrinsic to the concept of a palace, such as its large scale, the number of rooms, the variety of activities which it houses and the ostentation of luxury and wealth.
In the historical study of palaces, we recognised the impossibility of defining them through a specific typology. This is because their architecture was inherited from temples, whose different shapes are linked to how worship and ceremonies are performed. It is, therefore, possible to deduce that if there are changes in the behaviour of believers, this will generate new architectural needs.
In the same way that architecture as a discipline has the potential to control how we carry out activities based on the qualities of the space in which they take place, our behaviours also have the power to transform space since cultural protocols configure the abstract medium on which organisations are designed and standards of normality are set up. The more generic and flexible these spaces are, the longer they will last and the more resilient they will be.
The agropalace carries out a transmutation of power through which it frees itself from the human being as the centre and takes all the entities of the ecosystem as sovereign, understanding cohabitation as the central condition for the survival of the planet and human beings as a species.
The greenhouse typology appears as an architectural solution capable of regulating the climatic conditions in those places where there was a need to cultivate but where the climate was not entirely suitable. Agropalaces can not only incorporate productive spaces but generate entire ecosystems, becoming an architecture for the non-human.
We take as a reference the Crystal Palace. The Crystal Palace was designed for the London International Exhibition in 1851 by Joseph Paxton. The internal differentiation of its structural module, the height and the shape of its roof generate architectural conditions that shape it as a humidity-controlling container, which allows us to use it as the basis of our agropalatial prototype.
Our prototype based on the Crystal Palace is designed at first as a sequence of cross-sections. Their variables are the width and height of the section, the height and width of the central nave, the slope of the roof, the number of vaults, an infrastructural channel that transports water and, finally, the encounter with the floor. Each of these variables contributes to regulating the amount of water that each biome requires.
The territorial organisation of the agropalaces must be a-hierarchical for coexistence to take place. Cooperation between agropalaces is required for the system to function. This cooperation is based on water exchange from one palace to the other. For this to occur, vermiform palaces must be in a topography prone to flooding, organised in the form of an archipelago.
The prototype project is located in the Baix Llobregat Agrarian Park in Barcelona, which is crossed by the Llobregat river ending up in a delta in the Mediterranean Sea. The Agrarian Park currently grows food to supply to all the neighbouring cities. Our main interest in the site lies in its hydrographic network which is fundamental in the construction of the archipelago since the position of each agropalace depends on its distance to its closest water source.
To create a humidity map to determine the location of the palaces on the territory we use a generative adversarial network (GAN). A GAN is a type of AI in which systems can make correlations between variables, classify data and detect differences and errors between them through the study of algorithms. Their performance is improved as they are supplied with more data.
The GAN is trained with a dataset of 6000 images, each of them containing 4 channels of information in the form of coloured zones. Each channel represents the humidity of a specific biome. The position of the coloured zones is related to the distance to the water sources that each biome requires. The GAN analyses every pixel of the images to learn the patterns of the position of the channels and to create new possible location maps with emerging hybridisation between biomes.
The first four biomes are ocean, rainforest, tundra, and desert. Our choice for these extreme ecologies is related to the impact that global warming will have on them and the hypothesis that their hybridisation will produce less hostile and more habitable areas.
We conclude that the hybridisation made by AI is irreplaceable by human methods. As such, we consider AI part of the group of authors, even though a later curation of its production is carried out, constituting a post-anthropocentric thesis from its conception.
Due to the hybridisation, a gradient of nine biomes and their zones within the territory are recognised in the GAN outputs. These are, from wettest to driest: ocean, wetland, yunga, rainforest, forest, tundra, grassland, steppe, and desert. The wetter palaces will always be located at a shorter distance from the water supply points while the drier ones will be located closer to the transit networks. The GAN not only expands the range of a variety of biomes but also gives us unexpected organisations without losing respect for the rules previously established.
The chosen image is used as a floor plan and allows us to define the palatial limits, which are denoted by changes in colour.
The territory, initially flat, must become a differentiated topography so that the difference in the heights of the palaces eases access to water for those that require greater humidity.
The palaces are linear, but they contort to occupy their place without interrupting the adjoining palaces, following the central axis of the zone granted by the GAN.
This territorial organisation, a-hierarchical, longitudinal and twisted, forms two types of circulations: one aquatic and one dry. The aquatic palaces tend to form closed circuits, without specific arrival points. An idle circulation, unstructured, designed to admire the resulting landscape of canyons. The other, arid, runs through desertic palaces along its axis and joins the existing motorways in the Llobregat, crossing the Oasis.
The protocols of the post-Anthropocene must exist in a stereotomic architecture, a vast and massive territory, almost undifferentiated from the ground.
As mentioned above, our agropalatial prototype is designed as a sequence of cross-sections. Each section constitutes an envelope which formal characteristics are based on that of the Crystal Palace and modified concerning its need to hold water.
The determination of the interior spaces in each section depends on the fluxes of humidity necessary for generating the biome. The functional spaces are the result of the remaining space between the steam columns, the number of points where condensed water overflows towards the vaults, and the size of the central circulation channel.
The variation in organisation according to the needs of each biome creates different amounts of functional spaces, of different sizes and shapes, allowing the protocols to take place inside of them.
The interstices where the fluxes of humidity move are organised in such a way that the forces that travel through the surfaces of the functional spaces between them reach the ground on the sides of the palace, forming a system of structural frames.
The functional spaces in each cross-section are classified into three categories corresponding to the main protocols that take place inside of the agropalaces: production, housing and leisure.
The classification depends on the size, shape, distance to light and water of each functional space, predicting which one would be more convenient to house each protocol. Every cross-section contains at least one functional space of each kind.
These two-dimensional spaces are extruded, generating the “permanent” spaces, in which the activities are carried out. These form connections with the “permanent” spaces of the same category of the subsequent cross-section, forming “passage” spaces.
Thus, three unique, long, complex spaces – one for each protocol – run longitudinally through the palaces, in which activities are carried out in an interconnected and dynamic way. The conservation protocol – the biome itself – is the only non-sequential activity, since it is carried out in the interstice between the exterior envelope of the agropalace and the interior spaces.
The need for production has made cities and agricultural areas hyper-specialised devices, making their differences practically irreconcilable. However, we understand that this system is obsolete, which is why it is necessary to emphasise their deep connection and how indispensable they are to each other.
For this reason, agropalaces work through the articulation of different scales and programs, considering the three key pillars on which we must rely to build a new post-anthropocentric way of life – ecological conservation, agricultural production and human occupation – the latter prioritising leisure.
Protocol of Production
From currently available methods, we take hydroponic agriculture as the main means of production, together with aeroponic agriculture since both replace the terrestrial substrate with water rich in minerals.
The architectural organisation that shapes the agricultural protocol in the project is based on a central atrium that allows the water of the biome to condense and be redirected to the floodable platforms that surround it. In each biome, the density of the stalls, their depth, and the size of the central atrium vary in a linear gradient, ranging from algae and rice plantations to soybeans and fruit. The agricultural protocol in the agropalaces manages water passively, by surface condensation and gravity, generating a spiral distribution added to a central circulation that generates landscape while seeking to cultivate efficiently.
Protocol of Housing
In defining the needs for a House, Banham reduces it to an atmospheric situation, with no regard for its form. However, the dispossession of formal conditions allows us to modify the current housing protocol, through the ability to project a house whose shape is the result of passive climatic manipulation and the need to generate a variety of spatial organisations that do not restrict the type of social nuclei.
The spatial organisation of the house in the project is built through circulatory axes and rooms. The position of the circulatory axes and the number and size of the rooms vary depending on the biome, this time not based on humidity, but on the type of life that each ecological environment encourages. The height and width of the spaces also vary, generating the collision of rooms and thus allowing the formation of larger spaces or meta-rooms. The protocol of habitation in the agropalaces then allows a wide range of variation in which people are free to choose the form in which they wish to live, temporarily or permanently, individually or in groups.
Protocol of Leisure
Leisure is one of the essential activities of the post-Anthropocene because it frees human beings from their proletarian condition, characteristic of current capitalism, and connects them with the enjoyment of themselves and their surroundings. The leisure protocol in the thesis consists of a series of slabs with variable depths that constitute pools at different levels, interconnected by slides, which are to varying degrees twisted or straight, steep or shallow, and covered or uncovered.
The leisure protocol is based on the behaviour of water, which varies in each biome. The quantity, depth and position of the pools decrease in quantity the more desertic the biome that houses it is. In this way, water parks and dry staggered spaces are generated in which all kinds of games and sports are developed. In the agropalaces, contrary to being relegated to specific times and places, leisure becomes a form of existence itself.
Finally, to achieve coexistence, the architecture developed must be permeable. All the layers that contribute to the complexity of the project exchange fluids – mainly water – with the environment.
Water penetrates each of them, they use it to generate the desired ambient humidity for their biome and the excess then overflows on the roof. The system works sequentially, from the wettest to the driest biomes. Once the former palace overflows its residual water, the succeeding one can use it to its advantage until it eventually overflows again.
Inside every palace, a sequence of overflows on an intra-palatial scale is generated. Humidity enters the agropalace through its internal channel, where it evaporates and rises until it condenses on the surfaces of the functional organs and thus penetrates them to be used in different activities. The residuary water evaporates again until it overflows. The process consists of a cyclical system with constant recirculation.
The functional spaces’ envelopes have perforations in different sizes and positions to allow moisture to dissipate or condense as convenient. The overflowing quality of the system creates communication between the different scales of the architectural system, thus generating inter- and intra-palatial dependency.
Post-Anthropocentric Architecture: Conclusion
The agropalace understands coexistence as a necessary condition for the survival of the planet and human beings as a species. This new typology presents agriculture as the principal tool of empowerment and suggests a paradigm shift in which each society can define its policies for food production, distribution and consumption; meanwhile, it produces ecosystemic habitats with specific microclimatic qualities that allow the free development of all kinds of entities.
Biomatic Artefacts proposes an architecture whose forms do not interrupt the geological substrate but compose it, being part of the planetary ecology and simultaneously forming smaller-scale ecosystems within each palace and an autonomous ecosystem.
The protocols of today disappear to make room for the formation of a single para-protocol, since, contrary to being carried out in a single, invariable way, it exists because it has the quality of always being different, vast in spatial, temporal, and atmospheric variations. And in its wake, it generates a landscape of canyons and palaces that, in the interplay of reflections and translucency of water and glass, allows us to glimpse the ecological chaos of coexistence within.
We consider that the project lays the foundations for a continuation of ideas on agropalatial architecture and post-anthropocentric architecture, from which all kinds of new formal and material realities will come about.
The following paper was developed within the institutional framework of the School of Architecture and Urban Studies of Torcuato Di Tella University as a project thesis, with Lluis Ortega as full-time professor and Ciro Najle as thesis director.
 T. Morton, Hyperobjects: Philosophy and Ecology after the End of the World (Minnesota, USA: University of Minnesota Press, 2013).
 W. Steffen, P. Crutzen, J. McNeill, “The Anthropocene: Are Humans Now Overwhelming the Great Forces of Nature?”, AMBIO: A Journal of the Human Environment (2007), pp 614-621.
 T. Morton, Ecology Without Nature: Rethinking Environmental Aesthetics (Cambridge, USA: Harvard University Press, 2007).
 A. Reeser Lawrence, A. Schafer, “2 Architects, 10 Questions On Program, Rem Koolhaas + Bernard Tschumi” Praxis 8 (2010).
 C. Najle, The Generic Sublime (Barcelona, España: Actar, 2016).
 Set of base images with which the GAN trains by identifying patterns and thus learning their behaviours. In our case, the dataset is based on a set of possible biome location maps based on proximity to water sources and highways.
 R. Banham, F. Dallagret, “A Home Is Not a House”, Art in America, volumen 2 (1965) pp 70-79.
Interspecies Encounters and Performative Assemblages of Contamination
Our inner mental ecology has been known to be fundamental for the meaningful and complete success of the notion of ecology. Further demonstrated by the neurosciences, we have now assimilated the notion that we first empathise emotionally and physiologically with what surrounds us in a precognitive phase and only at a later time do we understand consciously the source of our aesthetic experience and, cognitively, its reason and meaning.
In order to investigate the concept of digital and material contaminations as a new way to conceptualise democratic design processes as modes of appropriation and negotiation of space, we have chosen to venture into the epistemological ecotone between aesthetics and cognition, examined through the concept of affect. It is within affects, in fact, that creativity emerges through perception and a cognitive approach to change and social action, “bridging aesthetics and political domain” through a series of encounters between different ecologies and their becoming.
What the affect theory speculates is that our “life potential comes from the way we can connect with others”, from our connectedness and its intensity, to the point that the ability itself to connect with others could be out of our direct control. It is a question of affective attunement, an emergent experience that becomes proto-political, and as any experience that works through instantaneous assessments of affect it becomes also strongly connected with notions of aesthetics and cognition. The paper examines how both aesthetics and cognition could be the instantiators of a change of paradigm within affectual and post-humanist approaches to the design of our cities and territories.
THE DIMENSIONS OF POST-HUMANIST AESTHETICS
Aesthetics can be defined according to its field of reference in slightly different ways: in neuroscience, aesthetics is the neural manifestation of a process articulated into sensations, meaning and emotions; in evolutionary biology, aesthetics is an adaptive system to environmental stimuli; in an ecological discourse, aesthetics is capacity to respond to the patterns which connect; in philosophy and specifically in the context of Object-Oriented Ontology, aesthetics is the root of all philosophy. Above all, regardless of the framework of reference, aesthetics fundamentally represents a form of knowledge, and as such, it is a very powerful and uncanny conceptual device.
The choice to connect the topic of ecology with aesthetics is not only related to the idea that aesthetics is primarily a form of knowledge and because “any ecologic discourse must be aesthetic as well as ethical in order to be meaningful”, but also because aesthetics has the power to attract affects and to convey difficult or ambiguous concepts, like those feelings of ambivalence that often come along with the ecological debate. As Morton states, the aesthetic experience “provides a model for the kind of coexistence ecological ethics and politics wants to achieve between humans and nonhumans […] as if you could somehow feel that un-feelability, in the aesthetic experience”. As a form of semiotic and experiential information exchange, the aesthetic experience is our primary source of genuine human understanding.
Neuroscientist Damasio demonstrates through a compelling series of scientific studies how emotions are essential to rational thinking and social behaviour. In addition, the embodied simulation theory teaches us that in a precognitive phase we first empathise emotionally and physiologically with what surrounds us and only at a later stage understand consciously the source of our aesthetic experience and, cognitively, its reason and meaning.
“Our capacity to understand the others and what the others materially accomplish does not depend exclusively on theoretical-linguistic competences, but it strongly depends on our socio-relational nature, of which corporeity constitutes the deepest and not further reducible structure. … In this sense, the aesthetic experience is a process on multiple levels which exceeds a purely visual analysis and leans on the visceral-motor and somatomotor resonation of whoever experiences it.”
In other words, the theory speculates that the same neural structures involved in our bodily experiences, our sensing, contribute to the conceptualisation of what we observe in the world around us.
Aesthetics, however, is no competence nor ability nor property exclusive to human nature, it only depends on the different sensing apparatus of each agency – or on what the proto-ecologist von Uexküll defined as the Umwelt, a specific model of the world corresponding to a given creature’s sensorium. Being aware of this aesthetic “perceptual reciprocity”, of this condition of mutual affects towards the environment, opens up new perspectives of solidarities where multiple agencies, each one living through multiple temporalities and with their own “way of worlding”, participate in the remaking of the planet through their patterns of growth and reproduction, their polyarchic assemblages, their territories of action and their landscapes of affects. In fact, we need to acknowledge that the environment is constituted by an ecology of different forms of intelligence where humans are just one form of biochemical intensity.
This expanded notion of agency is further enriched by Bennett’s vital materialism, which by ascribing to non-living systems their own trajectories and potentials, defines a multidimensional gradient that includes not only human and biological intelligences, but the natural and the artificial, raw matter and machinic intelligence, revealing opportunities of intersection, contamination, and collaboration. Her thought is about the need to recognise the vital agency of matter “as the alien quality of our own flesh”, and a part of that “Parliament of Things” or “Vascularised Collective” mentioned by Latour in his Actor Network Theory.
This radical understanding of agency as a confederation of human and nonhuman elements, biological and artificial entities, leads to some critical questions regarding equality, accountability and moral responsibility. As a form of rhizomatic Animism, it aims to reclaim and honour the mesh of connections and “assemblages that generate metamorphic transformation in the capacity to affect and be affected – and also to feel, think, and imagine”. And it is this capacity to affect and be affected that once again emerges as the effectual and necessary catalyst for creation and change, as affects are implicated in all modes of experience as a dimension of becoming. They are located in a non-conscious “zone of indistinction” between action and thought, and they fully participate in cognitive processes.
This is a pervasive process that affects all scales of being singular and choral, from the mesoscale of large planetary processes down to the nano-mechanisms of molecular self-organisation, entailing a new worldly disposition towards the nature of being collective. And it’s precisely because of the trans-scalar and concurrent effects that this extended notion of agency produces while processing new interpretations and understandings of the world that, when considering its impact on ideas of the negotiation and democratisation of space, we should interrogate not only the larger mechanisms of collective sense and decision making, but the very processes of cognition, communication, and information exchange at its basis.
PERFORMING THE MANY VOICES
In recent publications, Hayles describes the idea of a cognitive non-conscious as the possibility for complex systems to perform functions that “if performed by conscious entities would be unquestionably called cognitive”. Drawing from artificial and biological examples, she further explores a series of complex, adaptive and intention-driven organisations that, performing within the domain of evolutionary dynamics, exhibit cognitive capacities operating at a level that is inaccessible to introspection. Within this context, when considering the relation between human cognition and the cognitive non-conscious, she explains, the human interpretation might enter algorithmic analysis at different stages, in a sort of dialogue that de-facto structures the potential outcomes of a hybrid cognitive process, where part of the interpretation might be outsourced to the cognitive non-conscious, in a process that intimately links the meaning of the information produced to the specific mechanisms and the context of the interpretation, opening multiple new opportunities for the interpretation of ambiguous information.
Indeed, the argument about the potential and the perils of automation for decision-making is as relevant as it is controversial today. Parisi is significantly more critical regarding the current practices of human-machine collaboration, warning of the dangers of granular machine-generated content amplifying existing bias, or worse, being redirected for a purpose not pre-known. “Even if algorithms perform non-conscious intelligence, it does not mean that they act mindlessly”, she argues. Building on Hayles’ argument, she further elaborates that while it is not possible to argue that cognition performed by non-conscious entities is coherent and able to link the past and the present in causal connection, it is possible for non-conscious cognition to expose “temporal lapses that are not immediately accessible to conscious human cognition”. This is a process that sees algorithms not just adapting passively to the data provided but establishing new patterns of meaning to form coevolutionary cognitive infrastructures that, based on the idea of indeterminacy as a model for automated and hybrid cognition, avoid the primary level of feedback based on prescriptive outcomes and incorporate parallelism of learning and processing.
These arguments acquire a particular relevance if further considered in combination with the theory of information expressed by Simondon, which, formulated as an antagonist argument to Shannon’s cybernetic theory of communication, argues that information is never found, but is always expressed through a process of individuation of the system, as the result of the tensions between the realities that compose the system itself, as the very notion that clarifies the modes through which these realities might become a system in the first instance. This is a process that, by drawing on Simondon’s notion of individuation as the process of social becoming that leads to the formation of the collective subject – the transindividual – becomes inherently metastatic as it emerges from the tension between the sensorial abilities of the system and its tropism.
As such, Simondon’s notion of transindividuality constitutes the basis for a radical reimagination of the process of becoming collective and building collective knowledge, and through its intersection with the speculative opportunities inherent in ideas of tropistic material computation, it also offers the potential for an emergent rearticulation of collective sense and decision making, ultimately offering a protocol towards the exploration of the material, technological and aesthetic dimensions of new post-human and pervasive forms of authorship.
Attempting to account for the multidimensional consequences of altering the creative processes as a result of the construction of collective authorship as an inherently transindividual practice, the points made above imply a series of strategies oriented toward the definition of emergent meaning potentially able to capture the weaker voices and signals. This includes a focus on the diverse sensual and affectual experience of the participants, the orientation towards procedural indeterminacy and the exploration of material intelligence.
Furthermore, if we consider them in their intersection with our initial idea of the environment as constituted by an ecology of different forms of intelligence – where the creation of aesthetic assemblages of collaborative agencies is intended as the entangled construction of space, time and value through the symbiosis of different forms of intelligence defined by open-endedness and inclusiveness – these ideas describe a new urban paradigm, where the notion and aesthetic language of single human authorship with intellectual ownership is substituted by the concept of a collective of humans and non-human ecologies that might recover the aesthetics’ real, fundamental meaning, as an ecological category.
It is with the acceptance of these mixtures of interchanges and crossings of energies, that we can finally observe the old notion of quality, as an essential and pure identity related to cathartic categories, giving way to a more diffused and impure version of itself; a definition of quality not so much related to pureness, homogeneity, uniformity and refinement, but rather to a more complex meaning of sophistication by collaboration, contamination and exploitation of multiple resonances and superimpositions.
As Lowenhaupt Tsing advocates, learning to look at multi-species worlds could lead to different types of production-based economies: “Purity is not an option if we want to pursue a meaningful, informed ecological discourse. We must acknowledge that contaminations are a form of liveable or bearable collaborations. ‘Survival’ requires liveable collaborations. Collaboration means working across differences which leads to contamination.”
These domains and agencies searched for across other species, other ecological intensities and other modes of cognition, and reconfigured through computational technology, respond to a different kind of beauty, a filthy one, a revolutionary one, and an ecologic one. One that, as Morton preaches, “must be fringed with some kind of slight disgust … A world of seduction and repulsion rather than authority”.
According to Guattari, such ecosophic aesthetic paradigms, these collective assemblages or abstract machines, working transversally on different levels of existence and collaboration, would organise a reinvention of social and ecological practices, offering opportunities for dialogues among different forms of ecological intensities. They would also instantiate processes that would give back to humanity a sense of responsibility, not just towards the planet and its living beings, but also towards that immaterial component which constitutes consciousness and knowledge. Such a change of perspective in terms of critical agency would inevitably bring along a change in what Jacque Rancière calls the distribution of the sensible – where sensible is understood as “perceptible or appreciable by the senses or by the mind”, in a definition that describes new forms of inclusion and exclusion of the human and non-human collectivity in the process of appropriation of reality. And since access to a different distribution of the sensible is “the political instrument par excellence against monopoly”, we should treasure it for its capacity to allow us, borrowing Thomas Saraceno’s words, “to tune in to the non-human voices that join ours in boundless connectivity canvases, … proposing the rhizomatic web of life, which highlights hybridisms between one species and another and between species and worlds”. This is a process that describes new trajectories for new forms of institutions where we shall consider not just individual democracy, but a democracy extended to other species, talking to us through the language of the machines.
TOWARDS CO-CREATIVE AFFECTUAL PRACTICES
Along these trajectories, when approaching world and space-making strategies, design processes are translated into an “entangled” construction of space, time, value, and resources, which are critically defined by the very processes of their formation. In such a perspective, artificial intelligence has the potential to become the enabler, the instantiator of a new wider democratic process potentially able to disrupt existing power structures, giving a voice to what currently has none: all the non-conscious agencies separate from humankind or its direct will.
This is a new form of authorship which translates the question to the final user, so that the inquiry is not so much what the user wants from the environment but what can the user do for the environment, an idea that reverts the role of the final user from consumer to service provider. Such a form of authorship takes place in a symbiosis of computational and non-computational forms of thinking, as a hybrid of the diverse modes of cognition, resulting in a new type of synthetic ecology: the one that the designer enables.
In such a context, digital design platforms work as co-evolutionary cognitive infrastructures dealing with an amalgamation of different types of resource thinking: the thinking coming from the machines, the thinking coming from human participants, and the one converging from other ecological intensities. This is a type of transindividual subjectivity, that, formed as an ecology of diverse forms of cognition, is choral, decentralised, and inclusive, and has the capacity of being able to transmit tacit or informed knowledge exposing new models of democratic collective decision- and sense-making. In this process, all the participating forms of cognition have the potential to learn from each other and to compose unexpected dialogues and collective knowledge – what we call “interfaces [i/f], physical/virtual devices, a platform, enabling communications among entities of different kinds each one with its own protocol of communication, knowledge, and values”. This is an approach to collective creation that, drawing on alternative ideas of communication and power between the participating agencies, maps the emergence and evolution of patterns of informed feedback, outlining the connections with ideas of learning and performative collaboration between human, synthetic and biological agencies. In the exploration of these new forms of authorship, designers face the challenge of orchestrating a process able to build fruitful associations between machine computation, genuine human understanding, and non-conscious cognitive agencies – a challenge that should be taken as an opportunity to construct open processes of self-reflection and learning.
The resulting Transindividualities, which are digital participatory scholarships to ecological and post-humanist theory, create the potential for the affirmation of novel mediated narratives, which, by challenging the responsibility of authorship, bring along a new definition of the Human and the need to reframe the question of the design of our cities and territories towards a Pervasive Affectual Urbanism, which points toward the urge of new ethos and new aesthetics.
The challenge will be perhaps best approached by objecting to the idea that the designer is exclusively and ultimately responsible for the design process, and by sustaining the hypothesis that the symbiosis between all the different types of ecologies inhabiting the space could welcome all sorts of different agents through a creative process that embraces indeterminacy. It will be about the belief that open-endedness, contamination, interaction, machine learning and genuine human understanding are not so much about consensus, but about layering and celebrating differences to best use all of them as resources toward the participatory project of space-making. It will be about praising quality as sophistication, by acceptance, negotiation, exploitation and rhizomatic contaminations of multiple resonances and superimpositions, where the value of the project will lie in the exchange of information which is not merely exchanged, but used to create again.
 F. Guattari, The Three Ecologies (London: The Athlone Press, 1987).
 A. Damasio, Descartes’ Error: Emotion, Reason, and the Human Brain (London: Putnam Pub Group, 1994).
V. Gallese, “Embodied Simulation: from Neurons to phenomenal experience”, in Phenomenology and the Cognitive Sciences 4 (Berlin: Springer, 2005), 23–48.
 B. Massumi, The Politics of affect (Cambridge: Polity Press, 2005).
 E. Manning, interviewed in B. Massumi, The Politics of affect (Cambridge: Polity Press, 2005), 135.
 B. Massumi, The Politics of affect (Cambridge: Polity Press, 2005).
 A. Chatterjee, The Aesthetic Brain: How We Evolved to Desire Beauty and Enjoy Art (Oxford: Oxford University Press, 2015).
 G. H. Orians, “An Ecological and Evolutionary Approach to Landscape Aesthetics”, in E. C. Penning-Rowsell, D. Lowenthal (Eds.), Landscape Meanings and Values (London: Allen and Unwin) 3–25.
 G. Bateson, Steps toward an ecology of mind (London: Wildwood house Limited, 1979).
 G. Harman, “Aesthetics as a First Philosophy: Levinas and the non-human”, Naked Punch 2012, http://www.nakedpunch.com/articles/147, accessed 3 Feb. 2020.
 F. Guattari, The Three Ecologies (London: The Athlone Press, 1987).
 T. Morton, All Art is Ecological (Milton Keynes: Penguin Books, Green Ideas, 2021).
 A. Damasio, Descartes’ Error: Emotion, Reason, and the Human Brain (London: Putnam Pub Group, 1994).
 V. Gallese, “Embodied Simulation: from Neurons to phenomenal experience”, in Phenomenology and the Cognitive Sciences 4 (Berlin: Springer, 2005), 23–48.
 J. Von Uexkull, A Foray into the Worlds of Animals and Humans (Minneapolis: University of Minnesota Press, 2010).
 D. Abram, The spell of the sensuous, Perception and language in a more-than-human world (New York: Vintage Books, 1997).
 B. Latour, Down to Earth. Politics in the New Climatic Regime (Cambridge, PolityPress, 2018).
 I. Di Carlo, “The Aesthetics of Sustainability. Systemic thinking and self-organization in the evolution of cities”, 2016, PhD thesis, University of Trento, IAAC, Barcelona, Spain.
 J. Bennett, Vibrant Matter. A political ecology of things (Durham N.C. and London: Duke University Press, 2010).
 B. Latour, We have never been modern (Cambridge: Harvard University Press, 1993).
 I. Stengers, “Reclaiming Animism”, e-flux, 2012, https://www.eflux.com/journal/36/61245/reclaiming-animism/, accessed 10 Oct. 2021.
 B. Massumi, Ontopower: War, Power, and the State of Perception (Durham N.C.: Duke University Press, 2015).
 K. N. Hayles, “Cognition Everywhere: The Rise of the Cognitive Non-conscious and the Costs of Consciousness”, New Literary History 45, 2, 2014.
 L. Parisi, “Reprogramming Decisionism”, e-flux, 2017, https://www.e-flux.com/journal/85/155472/reprogramming-decisionism.
 G. Simondon, L’individuazione psichica e collettiva, ed. and transl. P. Virno, (Rome: DeriveApprodi, 2001).
 A. Papeschi, “Transindividual Urbanism: Novel territories of digital participatory practice”, Proceedings from Space and Digital reality: Ideas, representations/applications and fabrications, 2019, 80-90.
 I. Di Carlo, “The Aesthetics of Sustainability. Systemic thinking and self-organization in the evolution of cities”, 2016, PhD thesis, University of Trento, IAAC, Barcelona, Spain.
 A. Lowenhaupt Tsing, The mushroom at the end of the world: on the possibility of life in Capitalist Ruins (New Jersey: Princeton Univ. Press, 2017).
 T. Morton, All Art is Ecological (Milton Keynes: Penguin Books, Green Ideas, 2021).
 F. Guattari, Chaosmosis. An ethico-aesthetic paradigm (Sydney: Power Publications, 1995).
 J. Ranciere, The Politics of Aesthetics: Politics and Aesthetics (New York: Continuum, 2014).
 T. Saraceno, “Aria”, Catalogue of the exhibition at Palazzo Strozzi Firenze (Venezia: Edizioni Marsilio, 2020).
 I. Di Carlo, “The Aesthetics of Sustainability. Systemic thinking and self-organization in the evolution of cities”, 2016, PhD thesis, University of Trento, IAAC, Barcelona, Spain.
 A. Papeschi, “Transindividual Urbanism: Novel territories of digital participatory practice”, Proceedings from Space and Digital reality: Ideas, representations/applications and fabrications, 2019, 80-90.
The purpose of this guide is to help authors ensure consistency with Prospectives Issues.
It includes the most contentious areas of spelling, punctuation and formatting. For more
general guidance on tone and style, please consult the UCL Author Guidelines and Content
Style Guide. Where this guide differs from UCL Author Guidelines or Content Style Guide, please
use this document. If helpful, you can also consult Issue 1 of Prospectives: https://journal.b-pro.org/issue/issue1
A design process consists of a conventionalised practice – a process of (personal) habits that have proven to be successful – combined with a quest for creative and innovative actions. As tasks within the field of architecture and urban design become more complex, professionals tend to specialise in one of many subsets, such as designing, modelling, engineering, managing, construction, etc. They use digital tools which are developed for these specialised tasks only. Therefore, paradoxically, automation and new algorithms in architecture and urbanism are primarily oriented to simplify tasks within subsets, rather than engaging with the complex challenges the field is facing. This fragmented landscape of digital technologies, together with the lack of proper data, hinders professionals’ and developers’ ability to investigate the full digital potential for architecture and urban design.  Today, while designers explore the aid that digital technologies can provide, it is mostly the conventionalised part of practice that is being automated to achieve a more efficient workflow. This position statement argues for a different approach: to overcome fragmentation and discuss the preconditions for truly coping with complexity in design – which is not a visual complexity, nor a complexity of form, but rather a complexity of intentions, performance and engagement, constituted in a large set of parameters. We will substantiate our statement with experience in practice, reflecting on the Retrofit Project: our goal to develop a smart tool that supports the design of energy neutral districts. 
So, can designers break free from the established fragmentation and compute more than technical rationale, regulations and socio-economic constraints? Can they also incorporate intentions of aesthetics, representation, culture and critical intelligence into an architectural algorithm? To do so, the focus of digital tools should shift from efficiency to good architecture. And to compute good architecture, there is a need to codify a designer’s evaluation system: a prescriptive method to navigate a design process by giving value to every design decision. This evaluation system ought to incorporate architectural liberty – and therein lies the biggest challenge: differentiating between where to apply conventionalised design decisions and where (and how) to be creative or inventive. Within a 5000-year-old profession, the permitted liberty for these creative acts has been defined elastically: while some treatises allow only a minimum of liberty for a designing architect, others will lean towards a maximum form of liberty to guarantee good architecture. 
A minor group of early adopters, such as Greg Lynn, Zaha Hadid Architects, and UN Studio, tried to tackle the field’s complexity using upcoming digital technologies, in the late ’90s early 2000s. They conveniently inferred their new style or signature architecture from these computational techniques. This inference, however, causes an instant divide between existing design currents and these avant-garde styles. The latter claim the notion of complexity – the justification for their computational techniques – lies mostly within the subset of form-giving, not covering the complexity of the field. This stylistic path is visible in, for example, Zaha Hadid Architects’ 2006 masterplan for Kartal-Pendik in Istanbul. The design thrives on binary decisions in 3D-modelling tool Maya, where it plays out a maximum of two parameters at once: the building block with inner court and the tower. The resulting plastic urban mesh looks novel and stylistically intriguing, yet produces no real urbanity and contains no intelligence on the level of the building type. This methodology does not generate knowledge on how well the proposed urban quarter (or constituent buildings) will perform on the level of, for example, costs, energy production and consumption, infrastructure, city utilities, diversity and health. The fluid mass still needs all conventional design operations to effectively turn it into a mixture of types, urban functions, and local identity. Arguably, the early adopters’ stylistic path avoided dealing with real complexity and remained close to simple automation. In doing so, while they promoted a digital turn, they might also have dug the foundations for today’s fragmentation in the field.
Ironically, to some extent Schumacher’s treatise – definitely the parts that promote parametricism as a style – reads as a cover-up of the shortcomings of parametric software; for example, the inability to produce local diversity and typological characteristics beyond formal plasticity.  Schumacher further rejects Vitruvius to prevent structural rationale from taking primacy, and he disavows composition, harmony and proportion as outdated variable communication structures to propose the “fluid space” as the new norm.  This only makes sense knowing that the alternative – a higher intelligence across the whole field of architecture and urban planning, such as codified data and machine learning algorithms – did not yet exist for the early adopters. Contemporary applications such as Delve or Hypar do make use of those intelligent algorithms, yet prioritise technical and economical parameters (e.g. daylight, density, costs) to market efficiency. 
Any endeavour to overcome the established fragmentation and simplified automation will ultimately find itself struggling with the question of what good architecture is. After all, even with large computational power at hand, the question remains: how to evaluate design decisions beyond the merely personal or functional, in a time where no unified design theory exists? In fact, the fragmented specialisation of today’s professionals has popularised the proclamation of efficiency. As a result, an efficiency driver (whether geared by controlling costs, management or resources) is often disguised as moral behaviour, as if its interest is good architecture first, and the profit and needs of beneficiaries only come second. If the added value of good architecture cannot be defined, the efficiency driver will continue to get the upper hand, eroding the architectural profession to an engineering and construction service providing calculations, permits and execution drawings.
It was inspiring to encounter Alessandro Bava’s Computational Tendencies on this matter:
The definition of what constitutes “good” architecture is, in fact, always at the center of architecture discourse, despite never finding a definite answer. Discourses around digital architecture have too often resolved the question of the “good” in architecture by escaping into the realm of taste or artistic judgment. 
Bava renders Serlio’s architectural treatise as an original evaluating system that attributes universal value, and revisits Rossi’s exalted rationalism to propose a merger of architecture’s scientific qualities with its artistic qualities. He aims to re-establish architecture’s habitat-forming abilities and prevent architecture from becoming an amalgam of reduced and fragmented services. However, Serlio’s treatise did not provide a fully codified and closed formal system, as it still includes the liberty of the architect.  Going through Serlio’s On Domestic Architecture, an emphasis is placed on ideal building types, mostly without context. Therefore, no consideration is given to how these types ought to be modified when they need to be fitted in less ideal configurations such as non-orthogonal grids. The books also remain ignorant of the exceptions: the corner-piece-type, or fitting-parts that mediate between buildings and squares on a higher level. This is not a cheap critique of Serlio’s work. It is an awareness one needs to have when revisiting Serlio’s work as a “proto-BIM system, one whose core values are not market availability or construction efficiency, but harmonic proportions”.  Arguably, it is the liberty, the modifications, and the exceptions that need to be codified, to reach beyond simplified automation, across fragmentation, and towards an architectural algorithm to assist designers.
This is easier said than done, otherwise the market would be flooded with design technologies by now. As with most design problems, the only way to solve them is by tackling them in practice. In 2021, the Design Sciences Hub, affiliated with the University of Antwerp, set up the Retrofit Project. The aim is to develop an application to test the feasibility of district developments. The solution will show an urban plan with an automatically generated function mix and optimized energetic and ecological footprint, for any given site and context. The project team collaborates with machine-learning experts and environmental engineers for the necessary interdisciplinary execution. Retrofit is currently in the proof-of-concept phase, which focuses on energy neutrality and will tackle urban health and carbon neutrality in the long run.
The problem of modifications and exceptions seems the easiest to examine, as it primarily translates into a challenge of computational power and coping with a multitude of parameters. However, these algorithms should be smart enough to select a specific range within the necessary modifications and exceptions to comply with the design task at hand. In this case, the algorithm should select the correct modifications and exceptions needed to integrate certain types into any given site within the Retrofit application. In other words, there is a need for an intelligent algorithm that can be fed a large number of types as input data to generate entirely new or appropriate building types. The catch resides within the word “intelligent”, as algorithms aren’t created intelligent, they are trained to reach a certain level of intelligence based on (1) codifiable theory and (2) relevant training sets of data. Inquiring into a variety of evaluation systems for architectural design that emerged over the last 40 years, Verbruggen revealed the impossibility of creating a closed theoretical framework, and uniquely relating this framework to a conventionalised evaluation system in practice.  As such, both the codifiable theory – a unified evaluation system that integrates scientific and artistic qualities into one set of rules – and the training set hardly exist in architecture and urban design. To complicate matters even more, today’s non-unification is itself often embraced as the precondition for good architecture [11-15]
And so, the liberty question emerges here once again: how can different types, their modifications and exceptions, including respective relationships with different contexts, be codified? It is easy to talk about codification, but much harder to implement it within a project. When different types are inserted into a database, how are the attributes defined? This is a task that proved to be very laborious and raised many new questions in the Retrofit project. Attributes will include shape and size, yet might also include levels of privacy, preferred material usage, degree of openness, average energetic performance, historic and social acceptance in specific areas, compatibility with different functions, etc. Which values define when and where a specific type is appropriate, and how are they weighed? Do architects alone fill up the database, and if so, which architect is qualified, and why? And when an AI application would examine existing typologies within our built environment, which of these examples should be considered good, and why? Can big data or IoT sensors help in data gathering? To truly take everything into account, how much data do we really need (e.g., a structure’s age and condition, social importance, usage, materials, history, etc.). Furthermore, when the Retrofit application runs on an artificially intelligent algorithm that is trained to think beyond the capabilities of a single architect, will the results diverge (too) much from what society is used to?
The many practical questions from the Retrofit Project show that defining the architect’s liberty is both the problem and holds the potential for digital technologies to tackle the true complexity of the field. Liberty is undeniably linked to the design process and, therefore, encoding a design process needs to (1) capture the architect’s evaluation system and (2) allow for targeted and smart data gathering. The evaluation system can then be coded into an algorithm, with the help of machine learning experts, and trained using the gathered data. Both the evaluation system and the necessary data rely heavily on the architect’s liberty. Because dealing with these liberties is a difficult task – perhaps the most difficult task in the age of digital architecture – many contemporary businesses and start-ups that claim to revolutionise the design process with innovative technologies might not revolutionise anything, because they opt for the easy route and avoid dealing with the liberty aspect. An architectural algorithm that does take the liberty aspect into account may provide designers with an artificial assistant to help tackle all complexities in the field while tapping into the full potential of today’s available computational power.
This could be the ultimate task we set ourselves at the DSH. Studying a large dataset of design processes, steps, and creative acts might reveal codifiable patterns that could be integrated into a unified and conventionalized evaluation system. This study would target large and diverse groups of designers and users in general, including their knowledge exchange with other involved professionals. Could such an integral evaluation system, combined with data gathering, finally offer the prospect of developing a truly architectural algorithm? Eventually, this too will encounter issues that require further study, such as deciding who to involve and how to wisely navigate between the highs and lows of the wisdom of crowds:  can we still trust the emerging patterns detected by machine learning algorithms to constitute proper architectural liberty and, thus, good architecture? We will proceed vigilantly, but we must explore this path to avoid further fragmentation, non-crucial automation, and the propagation of false complexity.
 N. Leach, Architecture in the Age of Artificial Intelligence: An Introduction for Architects (London; New York: Bloomsbury Visual Arts, 2021).
 The Design Sciences Hub [DSH] is a valorisation team of the Antwerp Valorisation Office. The DSH works closely with IDLab Antwerp for Machine Learning components and with the UAntwerp research group Energy and Materials in Infrastructure and Buildings [EMIB] to study energy neutrality within the Retrofit Project. Although the project will be led and executed by the University of Antwerp, the private industry is involved as well. Four real estate partners – Bopro, Immogra, Quares and Vooruitzicht – are financing and steering this project. So is the Beacon, maximizing the insights from digital technology companies. Also see: https://www.uantwerpen.be/en/projects/project-design-sciences-hub/projects/retrofit/
 H.W. Kruft, A History of Architectural Theory: from Vitruvius to the present (London; New York: Zwemmer Princeton Architectural Press, 1994).
 P. Schumacher, The Autopoiesis of Architecture: A New Framework for Architecture. Vol. 1 (Chichester: John Wiley & Sons Ltd, 2011). P. Schumacher, The Autopoiesis of Architecture: A New Agenda for Architecture. Vol. 2 (Chichester: John Wiley & Sons Ltd, 2012).
 Delve is a product of Sidewalk Labs, founded as Google’s urban innovation lab, becoming an Alphabet company in 2016. Hypar is a building generator application started by former Autodesk and Happold engineer Ian Keough. Also see www.hypar.io, www.sidewalklabs.com/delve.
 A. Bava, “Computational Tendencies”, In N. Axel, T. Geisler, N. Hirsch, & A. L. Rezende (Eds.), Exhibition catalogue of the 26th Biennial of Design Ljubljana. Slovenia (2020): e-flux Architecture and BIO26| Common Knowledge.
 H.W. Kruft, A History of Architectural Theory: from Vitruvius to the present (London; New York: Zwemmer Princeton Architectural Press, 1994).
 A. Bava, “Computational Tendencies”, In N. Axel, T. Geisler, N. Hirsch, & A. L. Rezende (Eds.), Exhibition catalogue of the 26th Biennial of Design Ljubljana. Slovenia (2020): e-flux Architecture and BIO26| Common Knowledge.
 S. Verbruggen, The Critical Residue: Creativity and Order in Architectural Design Theories 1972-2012 (2017).
 M. Gausa & S. Cros, Operative optimism (Barcelona: Actar, 2005)
 W. S. Saunders, The new architectural pragmatism: a Harvard design magazine reader. (Minneapolis: University of Minnesota Press, 2007).
 R. Somol & S. Whiting, Notes around the Doppler Effect and Other Moods of Modernism. (2002) In K. Sykes (Ed.), Constructing a New Agenda: Architectural Theory 1993-2009 (1st ed., pp. 188-203). (New York: Princeton Architectural Press).
 K. Sykes, Constructing a new agenda : architectural theory 1993-2009. (1st ed., New York: Princeton Architectural Press, 2010).
 S. Whiting, (recorded in Delft, march 2006). The Projective, Judgment and Legibility: Lecture at the Projective Landscape Conference, organized by the TU Delft and the Stylos foundation.
 P. Mavrodiev & F. Schweitzer “Enhanced or distorted wisdom of crowds? An agent-based model of opinion formation under social influence”, Swarm Intelligence, 15(1-2), 31-46. doi:10.1007/s11721-021-00189-3 J. Surowiecki, The wisdom of crowds : why the many are smarter than the few. (London: Abacus, 2005).
The close integration of design with computational methods is not just transforming the relationships between architecture and engineering; it also contributes to reshaping modes of knowledge development. This paper critically probes some issues related to this paradigm shift and its consequences on architectural practice and self-awareness, looking at the potential of typical teaching approaches facing the digital revolution. The authors, who teach an architectural design studio together, coming from different backgrounds and research fields, probe the topic according to their respective vantage points.
Over the last few decades, a design agency has developed of using digital tools for the interactive generation of solutions by dynamically linking analytic and/or synthetic techniques.
The analytic techniques make use of simulation, of the capability to forecast certain aspects of building performance. While in conventional practice simulation usually plays a consulting role in the later stages of the design process, in the new forms of agency it works as a generative device from the earlier phases.
The synthetic techniques address, on the other hand, more organic, para-biologic concepts – for instance “emergence, self-organization and form-finding” – looking for “benefits derived from redundancy and differentiation and the capability to sustain multiple simultaneous functions”. 
Structures and their conception stand out as a part of architectural design where the digital impact shows its clearest consequences. Candela, Eiffel, Nervi and Torroja considered for instance that calculations have to go in parallel with intuitive understanding of the form: “The calculation of stresses”, writes Torroja, “can only serve to check and to correct the sizes of the structural members as conceived and proposed by the intuition of the designer”.  “In this fundamental phase of design”, Nervi adds, “the complex formulas and calculation methods of higher mathematics do not serve. What are essential, however, are rough evaluations based on simplified formulas, or the ability to break down a complex system into several elementary ones”.  At the time, the computational aspects were overridingly cumbersome; Frontón Recoletos required from Torroja one hundred and fifty-eight pages of calculations with approximate methods. Classical analytical procedures provided limited tools for simulation: “It was mandatory for the engineer to supplement his analyses with a great deal of judgment and intuition accumulated over years of experience. Empiricism played a great role in engineering design; while some general theories of mechanical behaviour were available, methods for applying them were still under development, and it was necessary to fall back upon approximation schemes and data taken from numerous tests and experiments”. 
After the epoch of Nervi and Torroja, research and practice have been deeply influenced by the combined actions of computation toward a unifying approach to the different theories in mechanics, thanks to exponential performance improvements in the hardware, as well as achievements in symbolic and matrix languages, and discretization methods (e.g., boundary and finite elements methods) implemented in software. At present, the wide availability of computational methods and tools can produce numerical simulations out of complex forms, with the expectation of providing a certain degree of knowledge and understanding of mechanics, energetics, fluids, and acoustics. The compelling possibilities of boundary or finite element methods, plus finite difference or volume methods, has produced a shift from science of construction pioneers’ awareness that not everything can be built,  to the “unprecedented morphology freedom” of the present.  Therefore, “We are limited in what we can build by what we are able to communicate. Many of the problems we now face”, as Hugh Whitehead of Foster and Partners points out, “are problems of language rather than technology. The experience of Swiss Re established successful procedures for communicating design through a geometry method statement”. 
“Parametric modelling”, Foster and Partners stated, “had a fundamental role in the design of the tower. The parametric 3D computer modelling process works like a conventional numerical spreadsheet. By storing the relationships between the various features of the design and treating these relationships like mathematical equations, it allows any element of the model to be changed and automatically regenerates the model in much the same way that a spreadsheet automatically recalculates any numerical changes. As such, the parametric model becomes a ‘living’ model – one that is constantly responsive to change – offering a degree of design flexibility not previously available. The same technology also allows curved surfaces to be ‘rationalized’ into flat panels, demystifying the structure and building components of highly complex geometric forms so they can be built economically and efficiently”. 
Of course, communication is here understood within a very specific part of the design process, mainly connected with fabrication issues and their optimisation, but it is a concept that involves many layered levels of meaning.  Curiously, this shift from the physical to the immaterial reminds us of the same step made by Leon Battista Alberti, who conceived design as a purely intellectual construct and was obsessed by its transmission from idea to built form without information decay.  Digital innovation promises to better connect the engineering process (focus on the object) with the wider reality (the architectural perspective), enabling design teams to deal with increasingly complex sets of variables. Freedom comes, however, with the disruption of the design toolbox, usually more defined by constraints than capabilities, so that the resulting wild fluctuations of effects seem increasingly disconnected from any cause. Design choices are therefore looking for multifaceted narrative support – and the “Gherkin”, with its combination of neo-functional-sustainable storytelling and metaphorical shape, turns out to be emblematic from this point of view too. 
Furthermore, extensive numerical simulations raise a question as to what extent they prove reliable, both because of their intrinsic functionality and the “black box” effect connected to the algorithmic devices. Those latter, especially in the latest applications of artificial intelligence such as neural networks, produce results through processes that remain obscure even to their designers, let alone less-aware users. Besides, the coupling of simulation with generative modelling through interactivity may not assist the designer in developing the understanding that, in several cases, (small) changes in the (coded) hypotheses can produce radically different solutions. Thus, the time spent in simulating alternatives can be more profitably spent working on different design hypotheses, and on architectural, technological and structural premises, perhaps with simpler computational models.
Are we entering an era of computational empiricism, as some authors maintain? 
Languages of innovation
Generative modelling, morphogenesis, parametric tooling, computational and performative design… all these apparatuses have brought methodological innovation into closer integration among different disciplines, bridging the gaps between fields. Modelling the project, the main common aim of this effort, has from the beginning leaned on logics and mathematics as a shared lingua franca.  Since the 1960s, applied mathematics has extended its applications through the formalisation process of information technology, which has developed the tools and the models beneficial for the purposes of science and technology. Information and communication technology puts into effect “the standardisation and automation of mathematical methods (and, as such, a reversal of the relationship of domination between pure mathematics and applied mathematics and, more generally, between theory and engineering)”. 
The redefinition of roles, between theories and techniques when applied to design, began in mathematics and physics with a metamorphosis of language,  with a shift towards symbolic languages that have gone beyond the mechanics of structures and the thermodynamics of buildings, subjecting it to automatic calculus, and finalising it in computation.  “Today, it is a widely held view that the advent of electronic computation has put an end to this semiempirical era of engineering technology: sophisticated mathematical models can now be constructed of some of the most complex physical phenomena and, given a sufficiently large computer budget, numerical results can be produced which are believed to give some indication of the response of the system under investigation”. 
The straightforward capability to model and simulate projects, supported by the evidence of results, has given confidence in the emerging computational tools, highlighting the dualism between the desire to make these devices usable for a wide range of practitioners, in a variety of cases and contexts, and the exigency of grounding bases for deeper understanding within a reflective practice. Moreover, the very nature of using digital tools urges designers to face an increasing risk of becoming “alienated workers” who, in Marxian terms, neither own their means of production in actuality – software companies lease their products and protect them against unauthorised modifications – nor, above all, conceptually, since their complex machinery requires a specifically dedicated expertise. Therefore, within the many questions this paradigm shift is raising about the redefinition of theories and practices and their mutual relationship, a main concern regards educational content and approaches, in terms of their ability to provide useful knowledge to future practitioners and aid their impact on society. In the architectural design field – which traditionally crossbreeds arts, applied sciences, and humanities in order to fulfil a broad role of mediation between needs and desires – this means dealing with an already contradictory pedagogic landscape in which ideologically opposite approaches (namely method-oriented and case-oriented pedagogies) overlap.
The specific of architectural design teaching does not escape this tension between methodological ambitions, nurtured by modern thinking and its quest for rationalisation, and the interplay between generations, languages and attitudes involved by learning through examples – even with its paradoxical side effects. One would expect in fact that a “positive” (according to Christopher Alexander), rule-based training should yield more open-ended outcomes than the “negative”, academic, disciplinary learning by copying.  But, on the one hand, the methodological approach implies an idea of linear control – towards optimisation and performance as well as in social and political terms – which reveals its origin in Enlightenment positivism. The Durandian apparatuses so widespread after World War II, with their proto-algorithmic design grammars, ended up accordingly with the reproduction of strict language genealogies. A similar trend seems to be emerging nowadays, in the convergence toward the same effective solutions in arts, sports, and whatever else, as a by-product of digital efficiency – which even the very technical camp is questioning. On the other hand, tinkering with the interpretation and application of examples makes possible the transmission of the many unspoken and unspeakable aspects connected to any learning endeavour. Getting closer to “good” examples – testing their potential according to specific situations – allows their inner quality to be grasped, reignited in different conditions, and finally transcended. Since forgetting requires something to be forgotten, Alexander is somehow right in framing this teaching attitude as “negative”: ironically, imitation provides the learning experience through which personal voices can emerge and thrive.
Turpin Bannister considered that in “an age of science”, architects “abandoned the scientific approach as detrimental to the pure art of design. On even the simplest question they acquiesced to their engineer and so-called experts”.  The pervasive penetration of computation in design would probably have met Bannister’s approval. The consequences and methodological implications are so far-reaching that they raise questions: how must education deal with the increased role of interactive computation in architectural design? And, more generally, with techno-science, its languages and methodologies?
Architectural design still relies on a “political” attitude, and mediation between the “two cultures”  is a fundamental asset of its disciplinary approach. Even though the unity of knowledge has disappeared with the advent of modern science, as Alberto Pérez-Gómez stated,  we ideally aspire to become like renaissance polymaths, mastering state-of-the-art skills in the most disparate fields. But in the long time that separates us from Brunelleschi and Alberti, the amount of knowledge required by the different aspects of the practice, even those which are specifically architectural, has grown exponentially, and trying to get a minimum of mastery over it would demand a lifelong commitment and extraordinary personal qualities. Digital prostheses promise to close the gap between the desire for control over the many facets of the design process and the real possibility of achieving it. Some consequences of the augmented agency provided by new information and communication technologies are already evident in the overlapping occurring in the expanded field of the arts, with protagonists from different backgrounds – visual arts or cinema for instance – working as architects or curators and vice versa.  The power of the digital to virtually replace those “experts”, to whom, according to Turpin Bannister, architects outsource their own choices, seems to act therefore as an evolutionary agent against overspecialisation, confirming the advantage Bucky Fuller attributed to the architect as the last generalist. 
However, without understanding and manipulating what happens within the black box of the algorithm, we still face the risk of being “designed” by the tools we put our trust in, going on to accept a subordinate position. Speaking machine, as John Maeda has pointed out,  is becoming necessary in order to contribute innovatively to any design endeavour. The well-known Asian American researcher, designer, artist and executive comes from a coding background, later supplemented with the study and practice of design and arts (along with business administration). His educational path and personal achievements indicate that such an integration of expertise is possible and desirable, even though his logical-mathematical grounding is likely the reason he mostly works with the immaterial, exploring media design and the so-called experience economy. Architectural schools are therefore facing the issue of if, when, and how to introduce coding skills into their already super-crammed syllabuses – from which, very often, visual arts, philosophy, law, storytelling and other much needed approaches and competencies are absent. One can argue that coding would provide young professionals with expertise they could immediately use in the job market, enabling them to better interact with contemporary work environments. On the other hand, a deeper perspective shows how the “resistance” of architectural specificity produced exceptional results in revolutionary times: academic education acted for the Modern masters as both a set of past, inconsistent practices to overcome and a background that enhanced the quality of their new language.
Digitalisation looks like a further step along the process of the specialisation of knowledge, which unfolded hand-in-hand with the development of sciences, techniques, and their languages. Since the dawn of the modern age, architects have often tried to bring together a unified body of knowledge and methodology; first around descriptive geometry, and then around geometry as a specific discipline which “gives form” to mathematics, statics and mechanics. “Geometry is the means, created by ourselves, whereby we perceive the external world and express the world within us. Geometry is the foundation”, Le Corbusier writes in the very first page of his Urbanisme, trying to keep pace with modernisation and establishing a new urban planning approach according to its supposed “exactitude”.  But while hard sciences and their technical application rely on regularity of results in stable experimental conditions, architects are still supposed to give different answers to the same question – or, more precisely, to always reframe architectural problems, questioning them in different ways.
Considering the volatility of the present situation, opening up and diversifying the educational offer seems a viable bet, more so than the attempt to formulate a difficult synthesis. Only by being exposed to the conflict between the selective, deterministic optimisation promise of code-wise design, and the dissipative, proliferating, unpredictable interpretation of cases can architects find their own, personal way to resolve it.
 S. Roudavski, “Towards Morphogenesis in Architecture”, International Journal of Architectural Computing, 3, 7 (2009) https://www.academia.edu/208933/Towards_Morphogenesis_in_Architecture (accessed 24 March 2021).
 E. T. Miret, J. J. Polivka and M. Polivka, Philosophy of Structures, (Berkeley: University of California Press, 1958), 331.
 P. L. Nervi, Aesthetics and Technology in Building (Cambridge, Mass.; London; Harvard University Press: Oxford University Press, 1966), 199.
 T. Oden, K.-J. Bathe, “A commentary on Computational Mechanics”, Applied Mechanics Reviews, 31, 8 (1978), 1055-1056.
 “We can now wonder whether any type of imaginary surface, is constructible. The answer is in the negative. So: how to choose and how to judge an imagined form?” E. T. Miret, J. J. Polivka and M. Polivka, Philosophy of Structures, (Berkeley: University of California Press, 1958) 78.
 M. Majowiecki, “The Free Form Design (FFD) in Steel Structural Architecture–Aesthetic Values and Reliability”, Steel Construction: Design and Research, 1, 1 (2008), 1.
 A. Menges, “Instrumental geometry”, Architectural Design, 76, 2 (2006), 46.
 Foster and Partners, “Modeling the Swiss Re Tower”, ArchitectureWeek, 238 (2005), http://www.architectureweek.com/2005/0504/tools_1-1.html (accessed 10 April 2022)
 “[Marjan] Colletti aptly quotes Deleuze stating: ‘The machine is always social before it is technical.’ The direct interaction between the designer and the equipment provides a feedback system of communication. He argues that the computer should ‘be regarded neither as abstract nor as machine’, but rather as an intraface.” C. Ahrens, “Digital Poetics, An Open Theory of Design-Research in Architecture”, The Journal of Architecture, 21, 2, (2016), 315; Deleuze’s passage is in G. Deleuze, C. Parnet, Dialogues (New York: Continuum International Publishing, 1987), 126-12; Colletti’s in M. Colletti, Digital Poetics, An Open Theory of Design-Research in Architecture (Farnham: Ashgate, 2013), 96.
 “We shall therefore first lay down, that the whole Art of Building consists in the Design, and in the Structure. The whole Force and Rule of the Design, consists in a right and exact adapting and joining together the Lines and Angles which compose and form the Face of the Building. It is the Property and Business of the Design to appoint to the Edifice and all its Parts their proper Places, determinate Number, just Proportion and beautiful Order; so that the whole Form of the Structure be proportionable. Nor has this Design any thing that makes it in its Nature inseparable from Matter; for we see that the same Design is in a Multitude of Buildings, which have all the same Form, and are exactly alike as to the Situation of their Parts and the Disposition of their Lines and Angles; and we can in our Thought and Imagination contrive perfect Forms of Buildings entirely separate from Matter, by settling and regulating in a certain Order, the Disposition and Conjunction of the Lines and Angles.” L. B. Alberti, The Ten Books of Architecture (London: Edward Owen, 1755 ), 25.
 A. Zaera-Polo, “30 St. Mary Axe: Form Isn’t Facile”, Log, 4 (2005).
 See – along with Oden, Bathe, and Majowiecki – Paul Humphreys, “Computational Empiricism”, Topics in the Foundation of Statistics, ed. by B. C. van Fraassen (Dordrecht: Springer, 1997) and P. Humphreys, Extending Ourselves: Computational Science, Empiricism, and Scientific Method. (New York: Oxford University Press, 2004).
 C Alexander, Notes on the Synthesis of Form (Cambridge, Mass.; London: Harvard University Press, 1964).
 J. Petitot, “Only Objectivity”, Casabella, 518, (1985), 36.
 E Benvenuto, An Introduction to the History of Structural Mechanics (New York, N.Y.: Springer-Verlag, 1991).
 M. Majowiecki, “The Free Form Design (FFD) in Steel Structural Architecture–Aesthetic Values and Reliability”, Steel Construction: Design and Research, 1, 1 (2008), 1.
 T. Oden, K.-J. Bathe, “A commentary on Computational Mechanics”, Applied Mechanics Reviews, 31, 8 (1978), 1056.
 “There are essentially two ways in which such education can operate, and they may be distinguished without difficulty. At one extreme we have a kind of teaching that relies on the novice’s very gradual exposure to the craft in question, on his ability to imitate by practice, on his response to sanctions, penalties, and reinforcing smiles and frowns. … The second kind of teaching tries, in some degree, to make the rules explicit. Here the novice learns much more rapidly, on the basis of general ‘principles’. The education becomes a formal one; it relies on instruction and on teachers who train their pupils, not just by pointing out mistakes, but by inculcating positive explicit rules.” C. Alexander, Notes on the Synthesis of Form (Cambridge, Mass.; London: Harvard University Press, 1964), 35.
 T. C. Bannister, “The Research Heritage of the Architectural Profession”, Journal of Architectural Education, 1, 10 (1947).
 C. P. Snow, The Two Cultures and the Scientific Revolution (Cambridge University Press, 1962).
 A. Pérez-Gómez, Architecture and the Crisis of Modern Science (Cambridge, Mass.: The MIT Press, 1983).
 “Artists after the Internet take on a role more closely aligned to that of the interpreter, transcriber, narrator, curator, architect.” A. Vierkant, The Image Object Post-Internet, http://jstchillin.org/artie/vierkant.html (accessed 21 September 2015). The artist Olafur Eliasson, for instance, started up his own architectural office (https://studiootherspaces.net/, accessed 30 March 2021), and the film director Wes Anderson authored the interior design of the Bar Luce, inside the Fondazione Prada in Milan.
 “Fuller … noted that species become extinct through overspecialization and that architects constitute the ‘last species of comprehensivists.’ The multidimensional synthesis at the heart of the field is the most invaluable asset, not just for thinking about the future of buildings but for thinking about the universe. Paradoxically, it is precisely when going beyond buildings that the figure of the architect becomes essential.” Mark Wigley, Buckminster Fuller Inc.: Architecture in the Age of Radio (Zürich: Lars Müller, 2015), 71.
 J. Maeda, How to Speak Machine: Laws of Design for a Digital Age (London: Penguin Business, 2019).
 Le Corbusier, The City of Tomorrow and its Planning (London: John Rocker, 1929 ), 1.
The remains of a virtual city – possibly the first of its kind – can be found on servers all over the world.1 Geocities was launched as a series of districts, alleyways, and neighbourhoods where its inhabitants could build their own webpages. For the first time, the internet was given a structure in a way that its audience could relate to it on a human scale. Today, around 650 gigabytes of Geocities’s data remain accessible thanks to archiving efforts that ensured the recovery of some of the 38 million individual websites that existed at the time of GeoCities’ final closure in 2009.    
GeoCities was first launched in 1994 by David Bohnett and Dick Altman as a web hosting service, allowing its users to store and manage their website files.  Its initial name, Beverly Hills Internet, already hinted at the creators’ intention to develop a neighbourhood of websites, which would later mature into a geography of cities. The service offered a free plan with a generous two megabytes of storage to all users, known as the homesteaders, who were asked to choose a neighbourhood to reside in.  All of the city’s inhabitants occupied a defined space, in a defined surrounding, where their homepages were arranged within neighbourhoods. Each cluster of pages was spatially close to those which shared similar content, while each neighbourhood was defined by the broader topic into which they fit. As such, the company created and thematically organised its web directories into six neighbourhoods, which included Colosseum, Hollywood, RodeoDrive, SunsetStrip, WallStreet and West Hollywood. New neighbourhoods, as well as their suburbs, were later added as the site grew, and became part of the members’ unique web address with a sequentially assigned URL “civic address” (e.g., “www.geocities.com/RodeoDrive/54”). Chat rooms and bulletin boards were added soon after, fostering rapid growth of the city.  Each neighbourhood had its own forum, live chat, and even a list of all the homesteaders who celebrated their birthday each day.
By December 1995, when it changed its name to GeoCities, Beverly Hills Internet had over 20,000 homesteaders and over 6 million page-views per month.  Within this expansive organisation of web page clusters, a seamless sense of proximity between those who shared similar ideas naturally led to human behaviours such as kinship and affection between them.
Neighbourhoods are intrinsic parts of our urban fabric and a self-evident manifestation of how the cities we live in are structured.  Yet, we still struggle to grasp a proper definition of their totality, given the complex layers within them. In 1926, progressive educator David Snedden defined the term neighbourhood as “those people who live within easy ‘hallooing’ distance”, illustrating it as a space where one can easily catch the attention of another. 
This essay will explore the notion of an algorithmic neighbourhood, one that reflects – and derives from – parts of a physically built, “hallooing” urban neighbourhood. The internet lexicon of today descends seamlessly from a long lineage of architectural and spatial terminologies, such as firewall, coding architecture, homepage, platform, address, path, room, and location, among many others. In the translation from a physical reality that is shaped within our Latourian “critical zone”, some of these terminologies have shifted in their meaning when applied to new forms of digital space.  A parallel “digital critical zone” is generated, within which these algorithmic neighbourhoods sit.
Neighbourhood as a site of kinship and proximity
The artisanal web built through GeoCities allowed “user-generated content”, which had not yet adorned itself with pompous names or revolutionary pretensions.  It proved that even before the invention of Web 2.0 – which was later aimed at implementing social-media profiles – the web was, above all, a story of human beings who interact with one another and discuss the subjects close to them through the means at hand.
Urban studies professor Looker defines the United States as a nation of neighbourhoods.  This essay expands on this exposure of the continental urban fabric by exploring the communities of algorithmic kinship that exist within GeoCities’ virtual borders. Similar to physically built neighbourhoods, GeoCities’ urban structure fostered kinship and affection among its inhabitants. PicketFence, for example, was built to allow residents to share tips and advice on ‘Home Improvement Techniques’. The more experienced ‘Home Improvement’ users became the neighbourhood’s go-to people for navigating daily issues, reinforcing a shared communal knowledge. 
West Hollywood, which was subdivided into “Gay, Lesbian, Bisexual, and Transgender topics”, is another example of such algorithmic kinship. This neighbourhood was a predecessor of today’s social-media spaces where users can gather and exchange (sometimes hidden or undisclosed) realities across communities. West Hollywood’s users could leave messages, sign a guestbook, and share contact information with one another. The neighbourhood gave people an opportunity to share similar experiences and daily struggles, form alliances with other communities, and tackle queer rights collectively. Moreover, West Hollywood fostered arenas of “block-level solidarity”, where “bonds and loyalties – whether as enacted on real-life pavements or as represented in stories, images, and speeches”, allowed connections between the intimate lives of users, their GeoCities pages, and the “city block”. 
Proximity and reciprocal kinship were thus a foundational feature of GeoCities’ design: individuals, together with their personal pages, were at the centre of the Internet. In contrast, today’s platforms and digital services are structured in such a nested way that proximity is sometimes inconceivable, and individuals are reduced to anonymous consumers of information. Today, the information communications technology industry (ICT) is at the centre of the Internet.  Social media platforms still provide virtual spaces that allow communities to gather and share content with one another, fostering a certain degree of human interaction. However, the very structure within which they operate is fundamentally different from the ones used in early platforms such as GeoCities. While before, the digital matter – text, images, links – was spatially placed onto the transparent structure of the webpage, and you could clearly see the location of a jpeg file within the HTML lines of code, now it all runs through opaque interfaces.  These perfect facades are quasi-impenetrable for users, and hide the “black boxes” where algorithms operate as instruments of measurements and perception.  As a counterpart to algorithmic neighbourhoods, Caroline Busta defines social-media platforms as a grand bazaar, “with lanes of kiosks, grouped roughly by trade, displaying representative works to passers-by. At the back of the mini-shop is a trap door with stairs leading to a sub-basement where deals can be done”.  This multi-layered opaque architecture of the bazaar illustrates the complex structure that currently governs social-media platforms. In contrast, the algorithmic neighbourhoods of GeoCities attempted to encourage a transparent vision of the modes of portraiture in the digital realm, and defined tools for users to relate directly to it.
Neighbourhood as a site constantly ‘under construction’
A digital archaeologist scavenging through GeoCities’ remains would come across a vast number of “under construction” signs strewn across the neighbourhood’s alleys, outlining its “work-in-progress” state. Surrounded by virtual scaffolding, the pages under construction were built, line after line of code, by the homesteaders, slowly undergoing organic changes and upgrades. Each individual page was constructed by its creator, from its foundations to its decorative elements, in the HTML format – the HyperText Markup Language. The coding language not only allowed users to build their pages from scratch, but also to introduce multimedia resources such as JPGs and GIFs. A page under construction implies that there was a process of creation, which aimed at an eventual final form. Similar to a construction site, the individual web page could be openly observed throughout its making, as it could be visited by GeoCities inhabitants at any moment in time. It was a facade yet to come; a page that was shaped by the algorithmic manipulation of its users as they added another ‘about me’ section, a ‘guestbook’ to be signed, or a photo gallery of low-res pictures – to fit within the 2 megabytes limit – portraying their personal lives.
Differently, the architecture of new forms of webpages and content aggregators is now conceived with an opaque algorithmic structure. Their virtual space is not one of proximity and distance based on intelligible parameters, but one of hierarchical appearance and disappearance based on unintelligible instruments of perception.  For instance, Google’s page-ranking algorithm mutates and evolves over time, leaving no traces behind, except the ones it uses to train itself. When presented with Google search results, users are faced with a series of temporary choices that are the result of a very intricate mechanism of automatic selection and classification. Vladan Joler defines algorithms as “instruments of measurements and perception”; thus, algorithmic architecture can be outlined by an operation of the more-than-human. Data collection and consumer profiling are the parameters upon which the current Internet is being built, instead of it being a conscious construction process carried out by its users.
While the architectural backdrop of a platform is constantly being redefined based on who is interacting with it, its facade – the interface – is pure and familiar. This interface which we constantly visit, however, obscures what’s beneath it. Even if it is a clear manifestation of rules, as it tells you what you can or cannot do, it does not reveal through which mechanisms it gathers and conveys information, nor how the user’s actions are exploited for profitable means. The algorithmic design of GeoCities, based on neighbourhood alliances, had not yet allowed for this opacity, avoiding instances of power structures, black boxes, and opaque interfaces. It also avoided entering the black hole of rhizomatic surveillance that now permeates the virtual realm.  
Algorithmic neighbourhoods can also help to expose the physical infrastructure hosting them. Similarly, to the opaqueness of interfaces, our built neighbourhoods are shaped by an underground infrastructure of fleshly cables and routers. Data centres, globally connected by a web of cables, host our digital selves, which wander through the unmeasurable geographies of the Internet. They are out of reach, transcending any geographical boundary, as they mirror the ubiquitous nature of algorithmic spaces. Cables and data centres are, in fact, the physical side of the Internet, its thickness on our planet. They are the physical neighbourhood mirroring the algorithmic one, hosting the latter through servers, cables, connections, and energy. The physical neighbourhood which creates the digital infrastructure is not, however, a direct reflection of the algorithmic one. It is instead expansive, ubiquitous, fragmented, and absent, as it is designed to operate under strict safety protocols and privacy regulations.
Neighbourhood as a site of civic participation and resistance
The “Haunting” protest represented a point of resistance towards the tendency of tech-giants to channel social traffic through a corporate digital platform ecosystem – a ubiquitous model in today’s internet.  The organized response by the homesteaders was only possible by the virtue of the very architecture of GeoCities. Neighbourhoods allowed a bottom-up response that could contrast the overarching corporate control put in place by Yahoo. It was a gathering that was empowered by proximity and affection, while it could exploit the temporary nature of the homepages’ construction as a medium for political change. In 2009, in response to the termination of GeoCities by Yahoo, new mechanisms of neighbourly rebuttal emerged. The German hosting provider JimdoWeb, for instance, attempted to host the nomad homesteaders by launching the Lifeboat for GeoCities webpage. Simultaneously, efforts of internet archivists started to meticulously archive each homepage of GeoCities in a countering act to preserve memory and gather residues of the city.
The archived remains of the virtual city stand as an alternative approach to the complexity and opaqueness of the algorithmic layering of contemporary web-hosting services, as much as they reveal the ‘trans-scalar’ infrastructure of the Internet.  These neighbourly entanglements help us make sense of the current digital “global village”, offering an entry point to analyse how it is being shaped by the effects of globalisation, market economies, and imprudent media.   Moreover, they display how the global village is being governed by algorithmic interdependencies, which in turn affect the architectural formations in both virtual and physical realities. 
 Archive Team. Archiveteam.org. https://wiki.archiveteam.org/index.php?title=Main_Page (accessed April 16, 2022).
 R. Vijgen. “The Deleted City”, http://www.deletedcity.net/, (2017)
 Restorativland, “The Geocities Gallery”, https://geocities.restorativland.org/, (accessed March 1, 2022).
 “OoCities”, https://www.oocities.org/#gsc.tab=0, (accessed March 1, 2022).
 O. Lialina & D. Espenschied, “One Terabyte of Kilobyte Age”, Rhizome.org. https://anthology.rhizome.org/one-terabyte-of-kilobyte-age, (accessed March 1, 2022).
 A.J. Kim, Community Building on the Web: Secret Strategies for Successful Online Communities (United Kingdom: Pearson Education, 2006).
 B. Sawyer, D Greely, Creating GeoCities Websites, (Cincinnati, Ohio: Muska & Lipman Pub, 1999) .
 C. Bassett, The arc and the machine: Narrative and new media, (Manchester: Manchester University Press, 2013).
 J. Jacobs, “The City: Some Myths about Diversity”, The death and life of great American cities, (New York: Random House, 1961).
 R. Sampson, “The Place of Context: A Theory and Strategy for Criminology’s Hard Problems”, Criminology 51 (The American Society of Criminology, 2013).
 B. Latour, Critical Zones: The Science and Politics of Landing on Earth, (Cambridge, MA: MIT Press, 2020).
 B. Sawyer, D Greely, Creating GeoCities Websites, (Cincinnati, Ohio: Muska & Lipman Pub, 1999).
 B. Looker, A Nation of Neighborhoods: Imagining Cities, Communities, and Democracy in Postwar America, (Chicago: The University of Chicago Press, 2015).
 C. Busta, “Losing Yourself in the Dark”. Open Secret, KW Institute for Contemporary Art, https://opensecret.kw-berlin.de/essays/losing-yourself-in-the-dark/, (accessed April 16, 2022).
 S.U. Noble, Algorithms of Oppression: How Search Engines Reinforce Racism,. (United States: NYU Press, 2018).
 V. Joler, “New Extractivism”, Open Secret, KW Institute for Contemporary Art, https://opensecret.kw-berlin.de/artwork/new-extractivism/, (accessed April 16, 2022).
 C. Busta, “Losing Yourself in the Dark”. Open Secret, KW Institute for Contemporary Art, https://opensecret.kw-berlin.de/essays/losing-yourself-in-the-dark/, (accessed April 16, 2022).
 V. Joler, “New Extractivism”, Open Secret, KW Institute for Contemporary Art, https://opensecret.kw-berlin.de/artwork/new-extractivism/, (accessed April 16, 2022).
 D. Savat, “(Dis)Connected: Deleuze’s Superject and the Internet”, International Handbook of Internet Research, 423–36 (Dordrecht: Springer, 2009).
 K.D. Haggerty, R. Ericson, “The Surveillant Assemblage”. British Journal of Sociology, 51, 4, 605-622, (United Kingdom: Wiley-Blackwell for the London School of Economics, 2000).
 J. Hu, “GeoCitizens fume over watermark”, CNet.com, https://www.cnet.com/tech/services-and-software/geocitizens-fume-over-watermark/ (accessed March 1, 2022).
 R. Ku, Cyberspace Law: Cases and Materials, (New York: Wolters Kluwer, 2016).
 C. Reynolds, B. Hallinan, “The haunting of GeoCities and the politics of access control on the early Web”, New Media & Society, (United States: SAGE Publishing, 2021).
 B McCullough, “Interview with David Bohnett, founder of GeoCities”. Internet History Podcast, http://www.internethistorypodcast.com/2015/05/david-bohnett-founder-of-geocities/, (accessed April 16, 2022).
 J. Van Dijck, T. Poell, M. De Waal, The Platform Society: Public Values in a Connective World, (Oxford: Oxford University Press, 2018).
 A. Jaque, Superpowers of Scale, (New York: Columbia University Press, 2020).
 M. McLuhan, The Gutenberg galaxy: the making of typographic man (Toronto: University of Toronto Press, 1962).
 T. Friedman, The World Is Flat: A Brief History of the Twenty-First Century, (New York: Farrar, Straus and Giroux, 2005).  M. McLuhan, The Gutenberg galaxy: the making of typographic man, (Toronto: University of Toronto Press, 1962).
Architecture has always explored the latest technological advances, causing changes in the way architects represent and conceive design solutions. Over the past decades, these changes were mostly due to, first, the integration of new digital design tools, such as Computer-Aided Design (CAD) and Building Information Modelling (BIM), which allowed the automation of paper-based design processes , and then, the adoption of computational design approaches, such as Algorithmic Design (AD), causing a more accentuated paradigm shift within the architectural practice.
AD is a design approach based on algorithms that has been gaining prominence in both architectural practice and theory [2,3] due to its greater design freedom and ability to automate repetitive design tasks, while facilitating design changes and the search for improved solutions. Its multiple advantages have therefore motivated a new generation of architects to increasingly adopt the programming environments behind their typical modelling tools, going “beyond the mouse, transcending the factory-set limitations of current 3D software” [3; p. 203]. Unfortunately, its algorithmic nature makes this approach highly abstract, deviating from the visual nature of human thinking, which is more attracted to graphical and concrete representations than to alphanumerical ones.
To approximate AD to the means of representation architects typically use and thereby make the most of its added value for the practice, we need to lower the existing comprehension barriers, which hinder its widespread adoption in the field. To that end, this research proposes a new approach to the representation of AD descriptions – the Algorithmic Representation Space (ARS) – that encompasses, in addition to the algorithm, its concrete outputs and the mechanisms that contribute to its understanding.
Algorithmic Representation Method and Design Paradigms
Despite the cutting-edge aura surrounding it, AD is a natural consequence of architects’ desire to automate modelling tasks. In this approach, the architect develops algorithms whose execution creates the digital design model  instead of manually modelling it using a digital design tool. Compared to traditional digital modelling processes, AD is advantageous in terms of precision, flexibility, automation, and ease of change, allowing architects to explore wider design spaces easily and quickly. Two AD paradigms currently predominate, the main difference between them lying in the way algorithms are represented: architects develop their algorithms either textually, according to the rules of a programming language, or visually, by selecting and connecting graphical entities in the form of graphs . In either case, the abstract nature of the medium hinders its comprehension.
Algorithms are everywhere and are a fundamental part of current technology. In fact, digital design tools have long supported AD, integrating programming environments of their own to allow users to automate design tasks and deal with more complex, unconventional design problems. Unfortunately, despite its advantages and potential to overcome traditional design possibilities, AD was slow to gain ground in the field, remaining, after almost sixty years, a niche approach. One of the main reasons is the fact that it requires architects to learn programming, which is an abstract task that is far from trivial. This is aggravated by the fact that, for decades, most tools have had their own programming language, which in most cases was limited and hard to use, as well as a programming environment providing little support for the development and comprehension of algorithmic descriptions. Examples include ArchiCAD’s GDL (1983); AutoCAD’s AutoLisp (1986) and Visual Lisp (2000); 3D Studio Max’s MAXscript (1997); and Rhinoceros 3D’s Rhino.Python (2011) and RhinoScript (2007).
To make AD more appealing to architects and approximate it to the visual nature of architectural design processes, visual-based AD environments have been released in the meantime. In these environments, text-based algorithmic descriptions are replaced by iconic elements that can be connected to each other in dataflow graphs . Generative Components (2003) is a pioneering example that inspired more recent ones such as Grasshopper (2007) and Dynamo (2011). These tools offer a database of pre-defined operations (components) that users can access by simply dragging an icon onto the canvas and providing it with input parameters. For standard tasks covered by existing components, this speeds up the modelling task considerably. Furthermore, since programs are represented by graph structures – with nodes describing the functions, and the wires connecting them describing the data that gets transferred between them – it is easy to see which parts of the algorithm are dependent upon others, and thus, where the changes are propagated to. However, this is only true for small algorithms, which are a rare find in visual-AD descriptions . Therefore, despite solving part of the existing problems – which explains the growing popularity of this paradigm in the community – others have emerged, such as its inability to deal with more complex and larger-scale AD solutions [5,8,9].
In sum, AD remains challenging for most architects and a far cry from the representation methods they typically use. Human comprehension relies on concrete instances to create mental models of complex concepts . Contrastingly, AD, either visual or textual, operates at a highly abstract level. This grants it its flexibility but also hinders its comprehension.
Algorithmic Abstractness Vs Model Concreteness
Abstraction can be regarded as the process of removing detail from a representation and keeping only the relevant features . Some authors believe abstraction improves productivity: it not only focuses on the “big idea” or problem to solve  but also triggers creative thinking due to its vagueness, ambiguity, and lack of clarity .
Abstraction in architecture can be traced back at least as far as classical antiquity. Architectural treatises, such as Vitruvius’ “Ten Books on Architecture” , are prime examples of abstract representations because they intend to convey not specific design instances, but rather design norms that are applicable to many design scenarios. However, the human brain is naturally more attracted to graphical explanations than textual ones [15–17], a tendency that is further accentuated in a field with a highly visual culture such as architecture. For that reason, even the referred treatises were eventually illustrated after the birth of the printing press .
The algorithmic nature of AD motivates designers to represent their ideas in an abstract manner, focusing on the concept and its formal definition. This sort of representation provides great flexibility to the design process, as a single expression of an idea can encompass a wide range of instances that match that idea, i.e., a design space. Contrariwise, most representation methods, including CAD and BIM, compel designers to rapidly narrow down their intentions towards one concrete instance, on account of the labour required to maintain separate representations for each viable alternative.
In sum, abstraction gives AD flexibility and the ability to solve complex problems, but it also makes it harder to understand. Abstraction is especially relevant when dealing with mathematical concepts, such as recursion or parametric shapes; nature-inspired processes, such as randomness; and performance-based design principles, such as design optimisation. It is also critical when developing and fabricating unconventional design solutions, whose geometric complexity requires a design method with a higher level of flexibility and accuracy. Sadly, these are also the hardest concepts to grasp without concrete instances and visual aid.
Nevertheless, the described comprehension barrier, apparently imposed by the abstract-concrete dichotomy, is more obvious when the AD descriptions are independent entities with little to no connection to the outcomes they produce. Figure 1 represents the current conception of AD: there is a parametric algorithm, representing a design space, which can generate a series of design models when specific parameters are provided. We propose to overthrow this notion by including the outcomes of the algorithm in the design process itself, changing the traditional flow of design creation to accommodate more design workflows and comprehension approaches.
Algorithmic Representation Space
AD descriptions have an abstract nature, which is part of the reason they prove so beneficial to the architectural design process. However, when it comes to comprehending an AD – i.e., creating a mental model of the design space it represents – this feature becomes a burden. Human cognition seems to rely heavily on the accumulation of concrete examples to form a more abstract picture . For this reason, we advocate that, for a better comprehension of an AD, the algorithms themselves do not suffice.
This research proposes a new way to represent algorithmic descriptions that aids the development and understanding of AD projects. Under the name of Algorithmic Representation Space (ARS), this concept encompasses not only the algorithm but also its outcomes and the mechanisms that allow for the understanding of the design space it represents. AD descriptions stand to benefit significantly from the concreteness of the outputs they generate, i.e., the digital models. If we consider the models as part of the AD representation, we reduce its level of abstraction and increase its understandability, approximating it to the visual nature of human understanding. Nevertheless, we must also smooth its integration in more traditional design workflows, helping architects who still develop their models manually in digital design tools or are forced to use pre-existing models. Accordingly, the proposed ARS also enables the use of already existing digital models as starting points to arrive at an algorithmic description.
There are two core elements in the ARS (Figure 2), the algorithm and the model. The algorithm represents a design space in a parametric abstract way, which makes the multiple design alternatives it represents difficult to perceive. Contrastingly, each model represents an instance of a design space in a static but concrete way. Combining the former’s flexibility with the latter’s perceptibility is therefore critical for the success of algorithmic representation. For conceptual reasons, the presented illustration of the ARS levels the two elements. Nevertheless, one must keep in mind that the algorithm can generate potentially infinite digital models, and the concept holds for all of them.
We consider two entry points into the ARS: programming and modelling. Each will allow architects to traverse the ARS; in the former case, from algorithm to model, by running the instructions in the algorithm to generate a model; and in the latter, from model to algorithm, by extracting an algorithmic description capable of generating the design instance and then refactoring that description to make it parametric as well. In either case, it is important the ARS contemplates the visualisation of these algorithm-model relationships. Therefore, we propose including techniques such as traceability in any ARS. In the following section, we will use a case study, the Reggio Emilia Train Station by Santiago Calatrava, to illustrate the ARS and each of the proposed principles.
The typical AD process entails the creation of a parametric description that abstractly defines a design space according to the boundaries set by the architect (Figure 3). The parametricity of this description, or the size of the design space it represents, varies greatly with the design intent and the way it is implemented (e.g., degrees of freedom, rules, and constraints). By instantiating the parameters in the algorithm, the architect specifies instances of the design space, whose visualisation can be achieved by generating them in a digital design tool, such as a CAD, BIM, or game engine (Figure 3 – running the algorithm). Figure 4 presents several variations of the Reggio Emilia station achieved by running the corresponding AD description with varying input parameters, namely with a different number of beams, different beam sizes, and different amplitudes and phases of the sinusoidal movement.
Given the flexibility of this approach, the process of developing AD descriptions tends to be a very dynamic one, with the architect repeatedly generating instances of the design to assess the impact of the changes made at each stage. Consciously or not, architects already work in a bidirectional iterative way when using AD. However, this workflow can also greatly benefit from a more obvious showcasing of the existing relations between algorithm and model. Traceability mechanisms allow precisely for the visual disclosure of these relations (i.e., which instruction/component generated which geometry), and several AD tools support them already.
AD is not meant to replace other design approaches but, instead, to interoperate with them. This interoperability is important, to take advantage of the investment made into those well-established representation methods such as CAD and BIM, especially for projects where digital models already exist or are still being produced. Therefore, the second entry point to the ARS is the conversion of an existing digital model of a design into an AD program. This might be necessary, for instance, when we wish to optimise it for new uses and/or to comply with new standards . This process entails crossing the ARS in the opposite direction to that described in the previous section (Figure 5).
To convert a digital model into an AD description, there are two main steps: extraction and refactoring. Extraction entails the automatic generation of instructions that can reproduce an exact copy of the model being extracted. The resulting AD description, however, is non-parametric and of difficult comprehension. This is where refactoring comes in [20,21], a technique that helps to improve the AD description, increasing its readability and parametricity. While the first task can be almost entirely automated, and is currently partially supported by some AD tools, the second part depends heavily on the architect’s design intent and, thus, will always be a joint effort between man and machine. In either case, it is important that the ARS adapts to the multiplicity of digital design tools and representation systems that architects often use during their design process. They can use, for instance, 3D modelling tools, such as CADs or game engines, to geometrically explore their designs more freely, or BIM tools to enrich the designs with construction information and to produce technical documentation.
Navigating the ARS
As mentioned in the previous section, there are two main elements in the ARS: algorithms abstractly describing design spaces and digital models representing concrete instances of those design spaces. Either one can be accessed from either end of the spectrum, i.e., by programming and running the algorithm to generate digital models, or by manually modelling designs and then converting them into an algorithm. To allow for this bidirectionality between the two sides, the ARS relies on three main mechanisms: (a) traceability, (b) extraction, and (c) refactoring. The first allows the system to expose the existing relationships between algorithm and model in a visual and interactive way for a better comprehension of the design intent. The latter two allow us to traverse the ARS from model to algorithm, a less common crossing but an essential one, nevertheless. The following sections describe these three mechanisms in detail.
For a proper comprehension of ADs, architects must construct a mental model of the design space, comprehending the impact each part of the algorithm has in each instance of the design space. To that end, a correlation must be ever present between the two core elements of the ARS – algorithm and model – matching the abstract representation with its concrete realisation. Traceability establishes relationships amongst the instructions that compose the algorithm and the corresponding geometries in the digital model. This is particularly relevant when dealing with complex designs, as it allows architects to understand which parts of the algorithm are responsible for generating which parts of the model.
With traceability, users can select parts of the algorithm or parts of the model and see the corresponding parts highlighted in the other end. Grasshopper for Rhinoceros 3D and Dynamo for Revit, two visual AD tools, present unidirectional traceability mechanisms from the algorithm to the model. Figure 6 shows this feature at play in Grasshopper: users select any component on the canvas and the corresponding geometry is highlighted in the visualised model.
Regarding bidirectional traceability, there are already visual AD tools that support it, such as Dassault Systèmes’ xGenerative Design tool (xGen) for Catia and Bentley’s Generative Components, as well as textual AD tools, such as Rosetta , Luna Moth , and Khepri . Figure 7 shows the example of Khepri, where the user selects either instructions in the algorithm or objects in the model and the corresponding part is highlighted in the model or algorithm, respectively. Programming In the Model (PIM) , a hybrid programming tool, offers traceability between the three existing interactive windows: one showing the model, another the visual AD description, and a third showing the equivalent textual AD description.
Unfortunately, traceability is a computationally intensive feature that hinders the tools’ performance with complex AD programs – especially model-to-algorithm traceability, which explains why some commercial visual-based AD tools avoid it. Those that provide it inevitably experience a decrease in performance as the model grows. All referred text-based and hybrid options are academic works, built and maintained as proof of concept and not as commercial tools, which explains their acceptance of the imposed trade-offs. A possible solution for this problem is to allow architects to decide when to use this feature and only switch it on when the support provided compensates for the computational overhead . In fact, traceability-on-demand is Khepri’s current approach to the problem.
Extraction is the automatic conversion of a digital model into an algorithm that can faithfully replicate it. Previous studies [27,28] focused on the generation of 3D models from architectural plans or on the conversion of CAD to BIM models, using heuristics and manipulation of geometric relations. Sadly, the result is not an AD description, but rather another model, albeit more complex and/or informed. One promising line of research is the use of the probabilistic and neural-based machine learning techniques (e.g., convolutional or recurrent neural networks) that address translation from images to textual descriptions,  but further research is needed to generate algorithmic descriptions.
The main problems with extracting a parametric algorithm lie, first, in the assumptions the system would need to make while reading a finished model: for instance, distinguishing whether two adjacent volumes are connected by chance or intentionally and, if the latter, deciding if such connection should constitute a parametric restriction of that model or not. Secondly, it is nearly impossible to devise a system that can consider the myriad of possible geometrical entities and semantics available in architectural modelling tools.
Some modelling tools that favour the VP paradigm avoid this problem by placing the responsibility on the designer from the very start, restricting the modelling workflow and forcing the designer to provide the missing information. In xGen and Generative Components, the 3D model and the visual algorithm are in sync, meaning changes made in either one are reflected in the other. PIM presents a similar approach, extending the conversion to the textual paradigm as well, although it was only tested with simple 2D examples.
In practice, these tools offer real-time conversion from the model to the algorithm. However, either solution requires the model to be parametric from the start. Every modelling operation available in these tools has a pre-set correspondence to a visual component, and designers must build their models following the structured parametric approach imposed by each tool, almost as if they were in fact constructing an algorithm but using a modelling interface. As such, the system is gathering the information it needs to build parametric relations from the very beginning. This explains why neither xGen, nor Generative Components, nor PIM, can take an existing model created in another modelling software or following other modelling rules and extract an algorithmic description from it.
This problem has also been addressed in the TP field and promising results have been achieved in the conversion of bi-dimensional shapes into algorithms [24,30]. However, further work is required to recognise 3D shapes, namely 3D shapes of varying semantics, since architects can use a myriad of digital design tools to produce their models, such as CADs, BIMs, or game engines. Figure 8 presents an ideal scenario, where the ARS is able to extract an algorithm that can generate an identical model to that being extracted.
In either case, even if we arrive at the extraction of the most common 3D elements any time soon, the resulting algorithm will only accurately represent the extracted model, and it will comprise a low-level program, which is very hard for humans to understand. To make the algorithm both understandable and parametric, it needs to be further transformed according to the design intent envisioned by the architect. Increasing the algorithm’s comprehension level and the design space it represents is the goal of refactoring.
Refactoring (or restructuring) is commonly defined as the process of improving the structure of an existing program without changing its semantics or external behaviour . There are already several semi-automatic refactoring tools  that help to improve the readability and maintenance of algorithmic descriptions and increase their efficiency and abstraction level. Refactoring is an essential follow-up to an extraction process, since the latter returns a non-parametric algorithm that is difficult to decipher.
Figure 9 shows an example of a refactoring process that could take place with the algorithm extracted in Figure 8. The extracted algorithm contains numerous instructions, each responsible for generating a beam between two spatial locations defined by XYZ coordinates. It is not difficult to infer the linear variations presented in the first and fourth highlighted columns, which correspond to the points’ X values. To infer the sinusoidal variation in the remaining values, however, more complex curve-fitting methods would have to be implemented .
In either case, refactoring tools seldom work alone, meaning that a lot of user input is required. This is because there is rarely a single correct way of structuring algorithms, and the user must choose which methods to implement in each case. Refactoring tools, beyond providing suggestions, guarantee that the replacements are made seamlessly and do not change the algorithm’s behaviour. When trying to increase parametric potential, even more input is required, since it is the architect who must decide the degrees of freedom shaping the design space.
In our example (Figure 9), the refactored algorithm shown below has a better structure and readability but is still in an infant state of parametricity. As a next stage, we could start by replacing the numerical values proposed by the refactoring tool with variable parameters to allow for more variations of the sinusoidal movement.
Discussion and Conclusion
Architecture is an ancient profession, and the means used to produce architectural entities have constantly changed, not only integrating the latest technological developments, but also responding to new design trends and representation needs. Architects have long adopted new techniques to improve the way they represent designs. However, while, for centuries, this caused gradual changes in the architectural design practice, with the more accentuated technological development witnessed since the 60s, these modifications have become more evident. The emergence of personal computers, followed by the massification of Computer-Aided Design (CAD) and Building Information Modelling (BIM) tools, allowed architects to automate their previously paper-based design processes , shaping the way they approached design issues . However, these tools did little to change the way designs were represented, only making their production more efficient. It did not take long for this scenario to rapidly evolve with the emergence of more powerful computational design paradigms, such as Algorithmic Design (AD). Despite being more abstract and thus less intuitive, this design representation method is more flexible and empowers architects’ creative processes.
Given its advantages for architectural design practice, AD should be a complement to the current means of representation. However, to make AD more appealing for a wider audience and allow architects to make the most of it, we must lower the existing barriers by approximating AD to the visual and concrete nature of architectural thinking. To that end, we proposed the Algorithmic Representation Space (ARS), a representation approach that aims to replace the current one-directional conception of AD (going from algorithms to digital models) with a bidirectional one that additionally allows architects to arrive at algorithms starting from digital models. Furthermore, the ARS encompasses as means of representation not only the algorithmic description but also the digital model that results from it, as well as the mechanisms that aid the comprehension of the design space it represents.
The proposed system is based on two fundamental elements – the algorithm and the digital model – and architects have two ways of arriving at them – programming and modelling. Considering the first case, programming, the ARS supports the development of algorithms and the subsequent visualisation of the design instances they represent by running the algorithm with different parameters. In the second case, modelling, the ARS supports the conversion of digital models into algorithms that reproduce them. The first scenario allows AD representations to benefit from the visual nature of digital design tools, reducing the innate abstraction of algorithms and obtaining concrete instances of the design space that are more perceptible to the human mind. The second case enables the conversion of a concrete representation of a design instance into an abstract representation of a design space, i.e., a parametric description that can generate possible variations of the original design, benefiting from algorithmic flexibility and expressiveness in future design tasks.
To allow for this bidirectionality, the ARS relies on three main mechanisms: (a) traceability, (b) extraction, and (c) refactoring. Traceability addresses the non-visual nature of the first process – programming – by displaying the relationships between the algorithm and the digital model. Extraction and refactoring address the complexity of the second process – going from model to algorithm – the former entailing the extraction of the algorithmic instructions that, when executed, generate the original design solution, and the latter solving the lack of parametricity and perceptibility of the extracted algorithms by helping architects restructure them. The result is a new representation paradigm with enough (1) expressiveness to successfully represent architectural design problems of varying complexities; (2) flexibility to parametrically manipulate the resulting representations; and (3) concreteness to easily and quickly comprehend the design space embraced.
The proposed ARS intends to motivate a more widespread adoption of AD representation methods. However, it is currently only a theoretical outline. To reach its goal, the proposed system must gain a practical character. As future work, we will focus on applying and evaluating the ARS in large-scale design scenarios, while retrieving user feedback from the experience.
This work was supported by national funds through Fundação para a Ciência e a Tecnologia (FCT) (references UIDB/50021/2020, PTDC/ART-DAQ/31061/2017) and PhD grants under contract of FCT (grant numbers SFRH/BD/128628/2017, DFA/BD/4682/2020).
 S. Abubakar and M. Mohammed; Halilu, “Digital Revolution and Architecture: Going Beyond Computer-Aided Architecture (CAD)”. In Proceedings of the Association of Architectural Educators in Nigeria (AARCHES) Conference (2012)., 1–19.
 R. Oxman, “Thinking difference: Theories and models of parametric design thinking”. Design Studies (2017), 1–36. DOI:http://doi.org/10.1016/j.destud.2017.06.001
 K. Terzidis, “Algorithmic Design: A Paradigm Shift in Architecture ?” In Proceedings of the 22nd Education and research in Computer Aided Architectural Design in Europe (eCAADe) Conference, Copenhagen, Denmark (2004), 201–207.
 I. Caetano, L. Santos, and A. Leitão, “Computational design in architecture: Defining parametric, generative, and algorithmic design.” Frontiers of Architectural Research 9, 2 (2020), 287–300. DOI:https://doi.org/10.1016/j.foar.2019.12.008
 P. Janssen, “Visual Dataflow Modelling: Some thoughts on complexity”. In Proceedings of the 32nd Education and research in Computer Aided Architectural Design in Europe (eCAADe) Conference, Newcastle upon Tyne, UK (2014), 305–314
 E. Lee and D. Messerschmitt, “Synchronous data flow”. Proceedings of the IEEE 75, 9 (1987), 1235–1245. DOI:https://doi.org/10.1109/PROC.1987.13876
 D. Davis, “Modelled on Software Engineering: Flexible Parametric Models in the Practice of Architecture”. PhD Dissertation, RMIT University (2013).
 A. Leitão and L. Santos, “Programming Languages for Generative Design: Visual or Textual?” In Proceedings of the 29th Education and research in Computer Aided Architectural Design in Europe (eCAADe) Conference, Ljubljana, Slovenia (2011),139–162.
 M Zboinska, “Hybrid CAD/E Platform Supporting Exploratory Architectural Design”. CAD Computer Aided Design 59, (2015), 64–84. DOI:https://doi.org/10.1016/j.cad.2014.08.029
 D. Rauch, P. Rein, S. Ramson, J. Lincke, and R. Hirschfeld, “Babylonian-style Programming: Design and Implementation of an Integration of Live Examples into General-purpose Source Code”. The Art, Science, and Engineering of Programming, 3, 3 (2019), 9:1-9:39. DOI:https://doi.org/10.22152/programming-journal.org/2019/3/9
 H. Abelson, G.J. Sussman, and J. Sussman (1st ed. 1985), Structure and Interpretation of Computer Programs (Cambridge, Massachusetts, and London, England: MIT Press, 1996) DOI:https://doi.org/10.1109/TASE.2008.40
 B. Cantrell and A. Mekies (Eds.), Codify: Parametric and Computational Design in Landscape Architecture. (Routledge, 2018). DOI:https://doi.org/10.1017/CBO9781107415324.004
 A. Al-Attili and M. Androulaki, “Architectural abstraction and representation”. In Proceedings of the 4th International Conference of the Arab Society for Computer Aided Architectural Design, Manama (Kingdom of Bahrain) (2009), 305–321.
 M. Vitruvius, The Ten Books on Architecture. (Cambridge & London, UK: Harvard University Press & Oxford University Press, 1914).
 K. Zhang, Visual languages and applications. (Springer Science + Business Media, 2007).
 N. Shu, 1986, “Visual Programming Languages: A Perspective and a Dimensional Analysis”. In Visual Languages. Management and Information Systems, SK. Chang, T. Ichikawa and P.A Ligomenides (eds.). (Boston, MA: Springer, 1986). DOI: https://doi.org/10.1007/978-1-4613-1805-7_2
 E. Do and M. Gross, “Thinking with Diagrams in Architectural Design”. Artificial Intelligence Review. 15, 1 (2001), 135–149. DOI:https://doi.org/10.1023/A:1006661524497
 M. Carpo, The Alphabet and the Algorithm. (Cambridge, Massachusetts: MIT Press, 2011).
 I. Caetano, G. Ilunga, C. Belém, R. Aguiar, S. Feist, F. Bastos, and A. Leitão, “Case Studies on the Integration of Algorithmic Design Processes in Traditional Design Workflows”. In Proceedings of the 23rd International Conference of the Association for Computer-Aided Architectural Design Research in Asia (CAADRIA), Hong Kong (2018), 129–138.
 M. Fowler, Refactoring: Improving the Design of Existing Code. (Reading, Massachusetts: Addison-Wesley Longman, 1999)
 T. Mens and T. Tourwe, “A survey of software refactoring”. IEEE Transactions on Software Engineering. 30, 2 (2004), 126–139. DOI:https://doi.org/10.1109/TSE.2004.1265817
 A. Leitão, J. Lopes, and L. Santos, “Illustrated Programming”. In Proceedings of the 34th Annual Conference of the Association for Computer Aided Design in Architecture (ACADIA), Los Angeles, California, USA (2014), 291–300.
 P. Alfaiate, I. Caetano, and A. Leitão, “Luna Moth Supporting Creativity in the Cloud”. In Proceedings of the 37th Annual Conference of the Association for Computer Aided Design in Architecture (ACADIA), Cambridge, MA (2017), 72–81.
 M. Sammer, A. Leitão, and I. Caetano, “From Visual Input to Visual Output in Textual Programming”. In Proceedings of the 24th International Conference of the Association for Computer-Aided Architectural Design Research in Asia (CAADRIA), Wellington, New Zealand (2019), 645–654.
 M. Maleki and R. Woodbury, “Programming in the Model: A new scripting interface for parametric CAD systems:”. In Proceedings of the Annual Conference of the Association for Computer Aided Design in Architecture (ACADIA), Cambridge, Canada (2013), 191–198.
 R. Castelo-Branco, A. Leitão, and C. Brás, “Program Comprehension for Live Algorithmic Design in Virtual Reality”. In Companion Proceedings of the 4th International Conference on the Art, Science, and Engineering of Programming (<Programming’20> Companion), ACM, New York, NY, USA, Porto, Portugal, (2020), 69–76. DOI:https://doi.org/10.1145/3397537.3398475
 L. Gimenez, J. Hippolyte, S. Robert, F. Suard, and K. Zreik, “Review: Reconstruction of 3D building information models from 2D scanned plans”. Journal of Building Engineering 2, (2015), 24–35. DOI:https://doi.org/10.1016/j.jobe.2015.04.002
 P. Janssen, K. Chen, and A. Mohanty, “Automated Generation of BIM Models”. In Proceedings of the 34th Education and research in Computer Aided Architectural Design in Europe (eCAADe) Conference, Oulu, Finland, (2016) 583–590.
 J. Donahue, L. Hendricks, M. Rohrbach, S. Venugopalan, S. Guadarrama, K. Saenko, and T. Darrell, “Long-Term Recurrent Convolutional Networks for Visual Recognition and Description”. IEEE Transactions on Pattern Analysis and Machine Intelligence. 39, 4 (2017), 677–691. DOI:https://doi.org/10.1109/TPAMI.2016.2599174
 A. Leitão and S. Garcia., “Reverse Algorithmic Design”. In Proceedings of Design Computing and Cognition (DCC’20) Conference, Georgia, Atlanta, USA (2021). p. 317–328. DOI: https://doi.org/10.1007/978-3-030-90625-2_18
 P. Mogensen and A. Riseth, “Optim: A mathematical optimization package for Julia”. Journal of Open Source Software. 3, 24 (2018), 615. DOI:https://doi.org/10.21105/joss.00615
 T. Kotnik, “Digital Architectural Design as Exploration of Computable Functions”. International Journal of Architectural Computing 8, 1 (2010), 1–16. DOI:https://doi.org/10.1260/1478-0722.214.171.124
A loyal companion to the breakthroughs of artificial intelligence is the fear of losing jobs due to a robotic takeover of the labour market. Mary L. Gray and Siddharth Suri’s research on ghost work unveiled another possible future, where a “last mile” requiring human intervention would always exist in the journey towards automation.  The so-called “paradox of the last mile” has been exerting impacts on the human labour market across the industrial age, recurringly re-organising itself when absorbing marginalised groups into its territory. These groups range from child labourers in factories, to the “human computer” women of NASA, to on-demand workers from Amazon Mechanical Turk (MTurk).  Yet their strenuous efforts are often rendered invisible behind the ostensibly neutral algorithmic form of the automation process, creating “ghost work”. 
Based on this concept of “the last mile”, this study intends to excavate how its paradox has influenced architectural authorship, especially during architecture’s encounters with digital revolutions. I will firstly contextualise “architectural authorship” and “the last mile” in previous studies. Then I will discuss the (dis)entanglements between “automation” and “digitalisation”. Following Antoine Picon and Nicholas Negroponte, I distinguish between the pre-information age, information age and post-information age before locating my arguments according to these three periods. Accordingly, I will study how Leon Battista Alberti, the Fun Palace, and mass-customised houses fail in the last mile of architectural digitalisation and how these failures affect architectural authorship. From these case studies, I challenge the dominant narrative of architectural authorship, either as divinity or total dissolution. In the end, I contend that it is imperative to conceive architectural authorship as relational and call for the involvement of multi-faceted agents in this post-information age.
Architectural Authorship in the Digital Age
The emergence of architects’ authorial status can be dated back to Alberti’s De re aedificatoria, which states that “the author’s original intentions” should be sustained throughout construction.  Yet at the same time, those architects should keep a distance from the construction process.  It not only marks the shift from the artisanal authorship of craftsmen to the intellectual authorship of architects but also begets the divide between the authorship of architectural designs and architectural end products.  However, this tradition can be problematic in the digital age, when multi-layered authorship becomes feasible with the advent of mass-collaboration software and digital customisation technologies. 
Based on this, Antoine Picon has argued that, despite attempts to include various actors by collaborative platforms such as BIM, architects have entered the Darwinian world of competition with engineers, constructors and existing monopolies, to maintain their prerogative authorship over the profession.  These challenges have brought about a shifting attention in the profession, from authorship as architects to ownership as entrepreneurs.  Yuan and Wang, on the other hand, call for a reconciliation of architectural authorship between regional traditions and technologies from a pragmatic perspective.  However, these accounts did not throw off the fetters of positioning architects as the centre of analysis. In the following article, I will introduce “the last mile”, a theory from the field of automation, to provide another perspective on the issues of architectural authorship.
“The Last Mile” as Method
The meaning of “the last mile” has changed several times throughout history. Metaphorically, it was used to indicate the distance between the status quo and the goal
, in various fields, such as movies, legal negotiations, and presidential campaigns.  It was first introduced in the technology industry as “the last mile” of telecommunication, on which one of the earliest traceable records was written in the late 1980s.  Afterwards, “the last mile” of logistics began to be widely used in the early 2000s, following the dot-com boom of the late 90s that fuelled discussions of B2C eCommerce.  However, in this article, I will use “the last mile” of automation, a concept from the recent “AI revolution” since 2010, to reconsider architectural authorship.  In this context, “the last mile” of automation refers to “the gap between what a person can do and what a computer can do”, as Gray and Suri defined in their book. 
I employ this theory to discuss architectural authorship for two purposes.
1. Understanding the paradox of automation can be of assistance in understanding how architectural authorship changes along with technological advancements. Pasquinelli and Joler suggest that “automation is a myth”, because machines have never entirely operated by themselves without human assistance, and might never do so.  Subsequently, here rises the paradox that “the desire to eliminate human labour always generates new tasks for humans” and this shortcoming “stretched across the industrial era”.  Despite being confined within the architectural profession, architectural authorship is subject to change in parallel with the alterations of labour tasks.
2. I contend that changes in denotations of “the last mile” signal turning points in both digital and architectural history. As Figure 1 suggests, in digital history, the implication of the last mile has changed from the transmission of data to the analysis of data, and then to automation based on data. The former change was in step with the arrival of the small-data environment in the 1990s and the latter corresponds with the leap towards the big-data environment around 2010.  In a similar fashion, after the increasing availability of personal computers after the 90s, the digital spline in architecture found formal expression and from around 2010 onwards, spirits of interactivity and mass-collaboration began to take their root in the design profession.  Therefore, revisiting the digital history of architecture from the angle of “the last mile” can not only provide alternative readings of architectural authorship in the past but can also be indicative of how the future might be influenced.
Between Automation and Digitalisation
Before elucidating how architectural authorship was changed by the arrival of the automated/digital age, it is imperative to distinguish two concepts mentioned in the previous section – automation and digitalisation. To begin with, although automation first came to use in the automotive industry in 1936 to describe “the automatic handling of parts”, what this phrase alludes to has long been rooted in history.  As Ekbia and Nardi define, automation essentially relates to labour-saving mechanisms that reduce the human burden by transferring it to machines in labour-requiring tasks, including both manual and cognitive tasks.  Despite its use in human history, it was not until the emergence of digital computers after WWII that its meaning became widely applicable.  The notion of computerised automation was put forward by computer scientist Michael Dertouzos in 1979, highlighting its potential for tailoring products on demand.  With respect to cognitive tasks, artificial intelligence that mimics human thinking is employed to tackle functions concerning “data processing, decision making, and organizational management”. 
Digitalisation, on the other hand, is a more recent concept engendered by the society of information in the late 19th century, according to Antoine Picon.  This period was later referred to as the Second Industrial Revolution, when mass-production was made possible by a series of innovations, including electrical power, automobiles, and the internal combustion engine. It triggered what Beniger called the “control revolution” – the volume of data exploded to the degree that it begot revolutions in information technology.  Crucial to this revolution was the invention of digital computing, which brought about a paradigm shift in the information society.  It has changed “the DNA of information” in the sense that, as Nicholas Negroponte suggests, “all media has become digital”, by converting information from atoms to bits.  In this sense, Negroponte distinguishes between the information age, which is based on economics of scale, and the post-information age, founded on personalisation. 
It can be observed that automation and digitalisation are intertwined in multiple ways. Firstly, had there been no advancement in automation during the Second Industrial Revolution, there would be no need to develop information technology, as data would have remained at a manageable level. Secondly, the advent of digital computers has further intermingled these two concepts to the extent that, in numerous cases, for something to be automated, it needs first to be digitalised, and vice versa. In the architectural field alone, examples of this can be found in cybernetics in architecture and planning, digital fabrication, smart materials, and so on. Hence, although these two terms are fundamentally different – most obviously, automation is affiliated with the process of input and output, and digitalisation relates to information media – the following analysis serves with no intention to differentiate between the two. Instead, I discuss “the last mile” in the context of reciprocity between these two concepts. After all, architecture itself is at the convergence point between material objects and media technologies. 
Leon Battista Alberti: Before the Information Age
Digitalisation efforts made by architects, however, appeared to come earlier than such attempts made in industrial settings of the late 19th century. This spirit can be traced back to Alberti’s insistence on identicality during information transmission, by compressing two-dimensional and three-dimensional information into digits – which is exemplified by Descriptio Urbis Romae and De statua.  In terms of architecture, as mentioned previously, he positions built architecture as an exact copy of architects’ intention.  This stance might be influenced by his views on painting. First, he maintains that all arts, including architecture, are subordinate to paintings, where “the architraves, the capitals, the bases, the columns, the pediments, and all other similar ornaments” came from.  Second, in his accounts, “the point is a sign” that can be seen by eyes, the line is joined by points, and the surface by lines.  As a result, the link between signs and architecture is established through paintings since architecture is derived from paintings and paintings from points/signs.
Furthermore, architecture can also be built according to the given signs. In Alberti’s words, “the whole art of buildings consists in the design (lineamenti), and in the structure”, and by lineamenti, he means the ability of architects to find “proper places, determinate numbers, just proportion and beautiful order” for their constructions.  It can be assumed that, if buildings are to be identical to their design, then, to begin with, there must be “determinate numbers” to convey architects’ visions by digital means – such as De statua (Fig. 2). Also, in translating the design into buildings, these numbers and proportions should be unbothered by any distortions as they are placed in actual places – places studied and measured by digital means, just like Descriptio Urbis Romae (Fig. 2).
Although the Albertian design process reflects the spirit of the mechanical age, insisting on the identicality of production, it can be argued that his pursuit of precise copying was also influenced by his pre-modern digital inventions being used to manage data.  Therefore, what signs/points mean to architecture for Alberti can be compared to what bits mean to information for Negroponte, as the latter is composed of the former and can be retrieved from the former. Ideally, this translation process can be achieved by means of digitalisation.
Yet it is obvious that the last mile for Alberti is vastly longer than that for Negroponte. As Giorgio Vasari noted in the case of Servite Church of the Annunziata, while Alberti’s drawings and models were employed for the construction of the rotunda, the result turned out to be unsatisfactory, and the arches of nine chapels are falling backwards from the tribune due to construction difficulties.  Also, in the loggia of the Via della Vigna Nuova, his initial plan to build semi-circular vaults was aborted because of the inability to fulfil this shape on-site.  These two cases suggest that the allographic design process – employing precise measurements and construction – which heralded the modern digital modelling software and 3D-printing technologies, was deeply problematic in Alberti’s time.
This problem was recognised by Alberti himself in his De re aedificatoria, when he wrote that to be “a wise man”, one cannot stop in the middle or at the end of one’s work and say, “I wish that were otherwise”.  In Alberti’s opinion, this problem can be offset by making “real models of wood and other substances”, as well as by following his instruction to “examine and compute the particulars and sum of your future expense, the size, height, thickness, number”, and so on.  While models can be completed without being exactly precise, architectural drawings should achieve the exactness measured “by the real compartments founded upon reason”.  According to these descriptions, the design process conceived by Alberti can be summarised as Figure 3.
If, as previously discussed, architecture and its context can be viewed as an assembly of points and signs, the Albertian design process can be compared to how these data are collected, analysed and judged until the process reaches the “good to print” point – the point when architects exit and construction begins. Nonetheless, what Vasari has unveiled is that the collection, analysis and execution of data can fail due to technological constraints, and this failure impedes architects from making a sensible judgement. Here, the so-called “technological constraints” are what I consider to be “the last mile” that can be found across the Albertian design process. As Vasari added, many of these technological limitations at that time were surmounted with the assistance of Salvestro Fancelli, who realised Alberti’s models and drawings, and a Florentine named Luca, who was responsible for the construction process.  Regardless of these efforts, Alberti remarked that only people involved in intellectual activities – especially mathematics and paintings – are architects; the opposite of craftsmen.  Subsequently, the challenges of confronting “the last mile” are removed from architects’ responsibilities through this ostensibly neutral design process, narrowing the scope of who is eligible to be called an architect. The marginalisation of artisanal activities, either those of model makers, draughtsmen or craftsmen, is consistent with attributing the laborious last mile of data collection, analysis and execution – measuring, model making, constructing – exclusively to their domain.
While the division of labour is necessary for architecture, as John Ruskin argued, it would be “degraded and dishonourable” if manual work were less valued than intellectual work.  For this reason, Ruskin praised Gothic architecture with respect to the freedom granted to craftsmen to execute their own talents.  Such freedom, however, can be expected if the last mile is narrowed to the extent that, through digitalisation/automation, people can be at the same time both architects and craftsmen. Or can it?
Fun Palace: At the Turn of the Information and Post-Information Age
Whilst the Albertian allographic mode of designing architecture has exerted a profound impact on architectural discipline due to subsequent changes to the ways architects have been trained, from the site to the academy, this ambition of separating design from buildings was not fulfilled, or even agreed upon among architects, in the second half of the 20th century.  Besides, the information age on the basis of scale had limited influences on architectural history, except for bringing about a new functional area – the control room.  Architecture’s initial encounters with the digital revolution after Alberti’s pre-modern technologies can be traced back to the 1960s, when architects envisaged futuristic cybernetic-oriented environments.  Different from Alberti’s emphasis on the identicality of information – the information per se – this time, the digitalisation and information in architecture convey a rather different message.
Gorden Pask defined cybernetics as “the field concerned with information flows in all media, including biological, mechanical, and even cosmological systems”.  By emphasising the flow of data – rather than the information per se – cybernetics distinguishes itself in two aspects. Firstly, it is characterised by attempts of reterritorialization – it breaks down the boundaries between biological organisms and machines, between observers and systems, and between observers, systems and their environments, during its different development phases – which are categorised respectively as first-order cybernetics (1943-1960), second-order cybernetics (1960-1985) and third-order cybernetics (1985-1996). 
Secondly, while data and information became secondary to their flow, catalysed by technologies and mixed realities, cybernetics is also typified by the construction of frameworks.  The so-called framework was initially perceived as a classifying system for all machines, and later, after computers were made more widely available and powerful, it began to be recognised as the computational process.  This thinking also leads to Stephen Wolfram’s assertion that the physical reality of the whole universe is generated by the computational process and is itself a computational process.  This is where the fundamental difference is between the Albertian paradigm and cybernetics, as the former is based on mathematical equations and the latter attempts to understand the world as a framework/computation.  Briefly, in cybernetics theory, information per se is subordinate to the flow of information and this flow can again be subsumed into the framework, which is later known as computational processes (Fig. 4).
In Cedric Price’s Fun Palace, this hierarchical order resulted in what Isozaki described as “erasing architecture into system” after its partial completion (Fig. 5).  Such an erasure of architecture was rooted in the conceptual process, since the cybernetics expert in charge of the Fun Palace was Gordon Pask, who founded his theory and practice on second-order cybernetics.  Especially so, as considering that one major feature of second-order cybernetics is what Maturana and Varela termed “allopoiesis” – a process of producing something other than the system’s original component – it is understandable that if the system is architecture, then it would generate something different than architecture.  In the case of the Fun Palace, it was presupposed that architecture is capable of generating social activities, and that architects can become social controllers.  More importantly, Cedric Price rejected all that is “designed” and instead only made sketches of indistinct elements, diagrams of forces, and functional programs, rather than architectural details.  All these ideas, highlighting the potential in regarding architecture as the framework of computing – in contrast to seeing architecture as information – rendered the system more pronounced and set architecture aside.
By rejecting architecture as pre-designed, Price and Littlewood strived to problematize the conventional paradigm of architectural authorship. They highlighted that the first and foremost quality of the space should be its informality, and that “with informality goes flexibility”.  This envisages user participation by rebuking fixed interventions by architects such as permanent structures or anchored teak benches.  In this regard, flexibility is no longer positioned as a trait of buildings but that of use, by encouraging users to appropriate the space.  As a result, it delineates a scenario of “the death of the author” in which buildings are no longer viewed as objects by architects, but as bodily experiences by users – architectural authorship is shared between architects and users. 
However, it would be questionable to claim the anonymity of architectural authorship – anonymous in the sense of “the death of the author” – based on an insignificant traditional architectural presence in this project, as Isozaki did.  To begin with, Isozaki himself has remarked that in its initial design, the Fun Palace would have been “bulky”, “heavy”, and “lacking in freedom”, indicating the deficiency of transportation and construction technologies at that time.  Apart from the last mile to construction, as Reyner Banham explained, if the Fun Palace’s vision of mass-participation is to be accomplished, three premises must be set – skilful technicians, computer technologies that ensure interactive experiences and programmable operations, and a secure source of electricity connecting to the state grid.  While the last two concerns are related to technological and infrastructural constraints, the need for technicians suggests that, despite its claim, this project is not a fully automated one. The necessary involvement of human factors to assist this supposedly automated machine can be further confirmed in Price and Littlewood’s accounts that “the movement of staff, piped services and escape routes” would be contained within “stanchions of the superstructure”.  Consequently, if architects can extend their authorship by translating elements of indeterminacy into architectural flexibility, and users can be involved by experiencing and appropriating the space, it would be problematic to leave the authorship of these technicians unacknowledged and confine them within service pipes. 
The authorship of the Fun Palace is further complicated when the content of its program is scrutinized. Price and Littlewood envisaged that people’s activities would feed into the system, and that decisions would be made according to this information.  During this feed-in and feedback process, human activities would be quantified and registered in a flow chart (Fig. 6).  However, the hand-written proposed list of activities in Figure 6 shows that human engagement is inseparable from the ostensibly automated flow chart. The arrows and lines mask human labours that are essential for observing, recognising, and classifying human activities. These tasks are the last mile of machine learning, which still requires heavy human participation even in the early 21st century.
For instance, when, in 2007, the artificial intelligence project ImageNet was developed to recognise and identify the main object in pictures, developers found it impossible to increase the system’s accuracy by developing AI alone (and only assisting it when it failed).  Finally, they improved the accuracy of ImageNet’s algorithms by finding a “gold standard” of labelling the object – not from the developments of AI itself, but by using 49,000 on-demand workers from the online outsourcing platform MTurk to perform the labelling process.  This example suggests that if the automation promised by the Fun Palace is to be achieved, it is likely to require more than just the involvement of architects, users, and technicians. In the time of the Fun Palace’s original conception, the attempt was not fulfilled due to the impotence of computing technologies. Yet if such an attempt was to be made in the 2020s, it is likely that architectural authorship would be shared among architects, users, technicians, and ghost workers from platforms such as MTurk.
Returning to the topic of cybernetics, whilst cybernetic theories tend to redefine territories of the architectural system by including what was previously the other parts of the system – machines, observers, adaptive environments – the example of the Fun Palace has shown that this process of blurring boundaries would not be possible without human assistance, at least initially. The flow of information between these spheres would require human interventions to make this process feasible and comprehensible because, in essence, “the information source of machine learning (whatever its name: input data, training data or just data) is always a representation of human skills, activities and behaviours, social production at large”. 
Houses of Mass-Customisation: In the Post-information Age
Although cybernetics theories have metaphorically or practically influenced architectural discourse in multiple ways, from Metabolism and Archigram to Negroponte and Cedric Price, such impact was diminished after the 1970s, in parallel with the near-total banishment of cybernetics as an independent discipline in the in the academia.  After a long hibernation during “the winter of artificial intelligence”, architecture’s next encounter with digital revolutions happened in the 1990s.  It was triggered by the increasing popularity and affordability of personal computers – contrary to the expectations of cybernetics engineers, who back in the 1960s dreamt that computers would increase both in power and size.  These distinctive material conditions led to the underlying difference between the second-order cybernetics in the 1960s and architecture’s first digital turn in the 1990s. I contend that this distinction can be explained by comparing Turing’s universal machine with Deleuze’s notion of the “objectile”.
As Stanley Mathews argued, the Fun Palace works in the same way as the universal machine.  The latter is a precursor of modern electronic computers, which can function as different devices – either as typewriters, drawing boards, or other machines – according to different codes they receive (Fig. 7).  Comparatively, “objectile” connotes a situation in which a series of variant objects is produced based on their shared algorithms (Fig. 8).  These products are so-called “non-standard series” whose key definition relates to their variance rather than form.83
While the universal machine seems to require more power to support its every change, an infinite one-dimensional tape on which its programmers can mark symbols of any instructions to claim its universality, non-standard production can operate on a smaller scale and under less demanding environments.  The emphasis on variance in non-standard production processes also indicates a shift of attention from the “process” underscored by second-order cybernetics towards the product of certain parametric models. When the latter is applied to architecture, the physical building regains its significance as the variable product.
However, it does not mean a total cut-off between cybernetics and non-standard production. Since human-machine interactions are crucial for customising according to users’ input, I maintain that mass-customisation reconnects architecture with first-order cybernetics whilst resisting the notion of chaos and complexity intrinsic in second-order cybernetics.
Such correlation can be justified by comparing two examples. First, the visionary project Flatwriter (1967) by the Hungarian architect Yona Friedman proposed a scenario in which users can choose their preferred apartment plan from several patterns of spatial configurations, locations, and orientations.  Based on their preferences, they would receive optimised feedback from the system (Fig. 9).  This optimisation process would consider issues concerning access to the building, comfortable environments, lighting, communication, and so on.  Given that it rejects chaos and uncertainty by adjusting users’ selections for certain patterns of order and layout, this user-computer interaction system is essentially an application of first-order cybernetics, as Yiannoudes argued.  Contemporary open-source architectural platforms are based on the same logic. As the founder of WikiHouse argued, since the target group of mass-customisation is the 99 per cent who are constantly overlooked by the normative production of buildings after the retreat of state intervention, designing “normal” environments for them is the primary concern – transgression and disorder should be set aside.  As Figure 10 illustrates, similarly to Flatwriter, in theory, WikiHouse would pre-set design rules and offer design proposals according to calculations of the parametric model.  These rules would follow a “LEGO-like system”, which produces designs by arranging and composing standard types or systems.  Both Flatwriter’s optimisation and WikiHouse’s “LEGO-like system” are pursuing design in accordance with patterns, and discouraging chaotic results.
Nevertheless, neither Flatwriter nor WikiHouse has achieved what is supposed to be an automatic process of using parametric models to generate a variety of designs. For Flatwriter, the last mile of automation could be ascribed to the unavailability of computers capable of performing calculations or processing images. For WikiHouse, the project has not yet fulfilled its promise of developing algorithms for design rules that resemble how the “LEGO blocks” are organised. Specifically, in the current stage, plans, components and structures of WikiHouse are designed in SketchUp by hand.  The flexibility granted to users is achieved by grouping plywood lumber into components and allowing users to duplicate them (Fig. 11). Admittedly, if users are proficient in Sketchup, they could possibly customise their WikiHouse on demand – but that would then go against the promise of democratising buildings through open-source platforms. 
Consequently, the last mile of automation again causes a conundrum of architectural authorship. Firstly, in both cases, never mind “the death of the author”, it appears that there is no author to be identified. One can argue that it signals a democratic spirit, anonymising the once Howard Roark-style architects and substituting them with a “creative common”. Nonetheless, it must be cautioned that such substitution takes time, and during this time, architects are obliged to be involved when automation fails. To democratise buildings is not to end architects’ authorship over architecture, but conceivably, for a long time, to be what Ratti and Claudel called “choral architects”, who are at the intersection of top-down and bottom-up, orchestrating the transition from the information age of scale to the post-information age of collaboration and interactivity.  Although projects with similar intentions of generating design and customising housing through parametric models – such as Intelligent City and Nabr – may prove to be more mature in their algorithmic process, architects are still required to coordinate across extensive sectors – clients’ inputs, design automation, prefabrication, logistics, and construction.  Architectural authorship in this sense is not definitive but relational, carrying multitudes of meanings and involving multiplicities of agents. 
In addition, it would be inaccurate to claim architectural authorship by the user, even though these projects all prioritise users’ opinions in the design process. By hailing first-order cybernetics while rejecting the second-order, advocating order while disapproving disorder, they risk the erasure of architectural authorship – just as those who play with LEGO do not have authorship over the brand, to extend the metaphor of the “LEGO-like system” in WikiHouse. Especially as the digital turn in terms of technology does not guarantee a cognitive turn in terms of thinking.  Assuming that the capitalist characteristics of production will not change, technological advancements are likely to be appropriated by corporate and state power, either by means of monopoly or censorship.
This erasure of human agency should be further elucidated in relation to the suppression of chaos in these systems. As Robin Evans explained, there are two types of methods to address chaos: (1) preventing humans from making chaos by organising humans; and (2) limiting the effects of chaotic environments by organising the system.  While Flatwriter and WikiHouse choose to conform according to the former at the expense of diminishing human agency, it is necessary to reinvite observers and chaos as an integral part of the system towards mass-customisation and mass-collaboration (Fig. 12).
For Walter Benjamin, “the angel of history” moves into the future with its face turned towards the past, where wreckages were piled upon wreckages.  For me, addressing the paradox of “the last mile” in the history of architectural digitalisation is this backward gaze that can possibly provide a different angle to look into the future.
This article mainly discussed three moments in architectural history when technology failed to live up to the expectation of full automation/digitalisation. Such failure is where “the last mile” lies. I employ “the last mile” as a perspective to scrutinize architectural authorship in these moments of digital revolutions. Before the information age, the Albertian notational system can be regarded as one of the earliest attempts to digitalise architecture. Alberti’s insistence on the identical copying between designers’ drawings and buildings resulted in the divide between architects as intellectuals and artisans as labourers. However, this allographic mode of architectural authorship was not widely accepted even into the late 20th century.
At the turn of the information age and post-information age, Cedric Price’s Fun Palace was another attempt made by architects to respond to the digital revolution in the post-war era. It was influenced by second-order cybernetics theories that focused on the flow of information and the computational process. Buildings were deemed only as a catalyst, and architectural authorship was shared between architects and users. Yet by examining how the Fun Palace failed in the last mile, I put forward the idea that this authorship should also be attributed to technicians and ghost workers assisting the computation processes behind the stage.
Finally, I analysed two case studies of open-source architectural platforms established for mass-customisation. By comparing Flatwriter of the cybernetics era and WikiHouse of the post-information age, I cautioned that both systems degrade architectural authorship into emptiness, by excluding users and discouraging acts of chaos. Also, by studying how these systems fail in the last mile, I position architects as “choral architects” who mediate between the information and post-information age. Subsequently, architectural authorship in the age of mass-customisation and mass-collaboration should be regarded as relational, involving actors from multiple positions.
- Mary L. Gray and Siddharth Suri, Ghost Work: How to Stop Silicon Valley from Building a New Global Underclass (New York: Houghton Mifflin Harcourt Publishing Company, 2019).
- Gray and Suri.
- Gray and Suri.
- Mario Carpo, The Alphabet and the Algorithm (London: The MIT Press, 2011), p. 22.
- Carpo, The Alphabet and the Algorithm, p. 22.
- Carpo, The Alphabet and the Algorithm, pp. 22–23.
- Mario Carpo, The Second Digital Turn: Design Beyond Intelligence (Cambridge, MA: The MIT Press, 2017), pp. 131, 140.
- Antoine Picon, ‘From Authorship to Ownership’, Architectural Design, 86.5 (2016), pp. 39–40.
- Picon, ‘From Authorship to Ownership’, pp. 39 & 41.
- Philip F. Yuan and Xiang Wang, ‘From Theory to Praxis: Digital Tools and the New Architectural Authorship’, Architectural Design, 88.6 (2018), 94–101 (p. 101) <https://doi.org/10.1002/ad.2371>.
- ‘“The Last Mile” An Exciting Play’, New Leader with Which Is Combined the American Appeal, 10.18 (1930), 6; Benjamin B Ferencz, ‘Defining Aggression–The Last Mile’, Columbia Journal of Transnational Law, 12.3 (1973), 430–63; John Osborne, ‘The Last Mile’, The New Republic (Pre-1988) (Washington, 1980), 8–9.
- Donald F Burnside, ‘Last-Mile Communications Alternatives’, Networking Management, 1 April 1988, 57-.
- Mikko Punakivi, Hannu Yrjölä, and Jan Holmström, ‘Solving the Last Mile Issue: Reception Box or Delivery Box?’, International Journal of Physical Distribution and Logistics Management, 31.6 (2001), 427–39 <https://doi.org/10.1108/09600030110399423>.
- Gray and Suri, p. 12.
- Gray and Suri, p. 12.
- Matteo Pasquinelli and Vladan Joler, ‘The Nooscope Manifested: AI as Instrument of Knowledge Extractivism’, 2020, pp. 1–23 (p. 19) <https://doi.org/10.1007/s00146-020-01097-6>.
- Gray and Suri, pp. 12 & 71.
- Carpo, The Second Digital Turn: Design Beyond Intelligence, pp. 9, 18 & 68.
- Carpo, The Second Digital Turn: Design Beyond Intelligence, pp. 5, 18 & 68.
- James Beniger, The Control Revolution: Technological and Economic Origins of the Information Society (London: Harvard University Press, 1986), p. 295.
- Hamid R. Ekbia and Bonnie Nardi, Heteromation, and Other Stories of Computing and Capitalism (Cambridge, Massachusetts: The MIT Press, 2017), p. 25.
-  Ekbia and Nardi, pp. 25-6.
-  Michael L. Dertouzos, ‘Individualized Automation’, in The Computer Age: A Twenty-Year View, ed. by Michael L. Dertouzos and Joel Moses, 4th edn (Cambridge, Massachusetts: The MIT Press, 1983), p. 52.
- Ekbia and Nardi, p. 26.
- Antoine Picon, Digital Culture in Architecture : An Introduction for the Design Professions (Basel: Birkhäuser, 2010), p. 16.
- Beniger, p. 433.
- Picon, Digital Culture in Architecture : An Introduction for the Design Professions, pp. 24–26.
- Nicholas Negroponte, Being Digital (New York: Vintage Books, 1995), pp. 11 & 16.
- Negroponte, pp. 163–64.
- Carpo, The Alphabet and the Algorithm, p. 12.
- Carpo, The Alphabet and the Algorithm, pp. 54–55.
- Carpo, The Alphabet and the Algorithm, p. 26.
- Leon Battista Alberti, On Painting, trans. by Rocco SiniSgalli (Cambridge: Cambridge University Press, 2011), p. 45.
- Alberti, On Painting, p. 23.
- Leon Battista Alberti, The Ten Books of Architecture (Toronto: Dover Publications, Inc, 1986), p. 1.
- Carpo, The Alphabet and the Algorithm, p. 27.
- ‘Architectural Intentions from Vitruvius to the Renaissance’ [online] <https://f12arch531project.fil es.wordpress.com/2012/10/xproulx-4.jpg>; ‘Alberti’s Diffinitore’ http://www.thesculptorsfuneral.com /episode-04-alberti-and-de-statua/7zf3hfxtgyps12r9igveuqa788ptgj [accessed 23 April 2021].
- Giorgio Vasari, The Lives of the Artists, trans. by Julia Conaway & Peter Bondanella (Oxford: Oxford University Press, 1998), p. 182.
- Vasari, p. 181.
- Alberti, The Ten Books of Architecture, p. 22.
- Alberti, The Ten Books of Architecture, p. 22.
- Alberti, The Ten Books of Architecture, p. 22.
- Vasari, p. 183.
- Mary Hollingsworth, ‘The Architect in Fifteenth-Century Florence’, Art History, 7.4 (1984), 385–410 (p. 396).
- Adrian Forty, Words and Buildings: A Vocabulary of Modern Architecture (New York: Thames & Hudson, 2000), p. 138.
- Forty, p. 138.
- Forty, p. 137; Carpo, The Alphabet and the Algorithm, p. 78.
- Picon, Digital Culture in Architecture : An Introduction for the Design Professions, p. 20.
- Mario Carpo, ‘Myth of the Digital’, Gta Papers, 2019, 1–16 (p. 3).
- N. Katherine Hayles, ‘Cybernetics’, in Critical Terms for Media Stuies, ed. by W.J.T. Mitchell and Mark B.N. Hansen (Chicago and London: The University of Chicago Press, 2010), p. 145.
- Hayles, p. 149.
- Hayles, pp. 149–50.
- Socrates Yiannoudes, Architecture and Adaptation: From Cybernetics to Tangible Computing (New York and London: Taylor & Francis, 2016), p. 11; Hayles, p. 150.
- Hayles, p. 150.
- Stephen Wolfram, A New Kind of Science (Champaign: Wolfram Media, Inc., 2002), pp. 1, 5 & 14.
- Arata Isozaki, ‘Erasing Architecture into the System’, in Re: CP, ed. by Cedric Price and Hans-Ulrich Obrist (Basel: Birkhäuser, 2003), pp. 25–47 (p. 35).
- Yiannoudes, p. 29.
- Yiannoudes, p. 14.
- Stanley Mathews, ‘The Fun Palace as Virtual Architecture: Cedric Price and the Practices of Indeterminacy’, Journal of Architectural Education, 59.3 (2006), 39–48 (p. 43); Yiannoudes, p. 26.
- Isozaki, p. 34; Yiannoudes, p. 50.
- Stanley Mathews, p. 47.
- Cedric Price and Joan Littlewood, ‘The Fun Palace’, The Drama Review, 12.3 (1968), 127–34 (p. 130).
- Price and Littlewood, p. 130.
- Forty, p. 148.
- Jonathan Hill, Actions of Architecture (London: Routledge, 2003), pp. 68–69.
- Isozaki, p. 34.
- Isozaki, p. 35.
- Reyner Banham, Megastructure: Urban Futures of the Recent Past (London: Thames and Hudson, 1972).
- Price and Littlewood, p. 133.
- Forty, pp. 142-8.
- Yiannoudes, p. 29.
- Yiannoudes, p. 31.
- Gray and Suri, pp. 33–34.
- Gray and Suri, p. 34.
- Cedric Price, Fun Palace Project (1961-1985), <https://www.cca.qc.ca/en/archives/380477/cedric-price-fonds/396839/projects/399301/fun-palace-project#fa-obj-309847> [accessed 25 April 2021].
- Pasquinelli and Joler, p. 19.
- Yiannoudes, p. 18; Carpo, ‘Myth of the Digital’, p. 11; Hayles, p. 145.
- Mario Carpo, ‘Myth of the Digital’, pp. 11–13.
- Carpo, ‘Myth of the Digital’, p. 13.
- Mathews, p. 42.
- Yiannoudes, p. 33.
- Carpo, The Alphabet and the Algorithm, p. 99.
- Carpo, The Alphabet and the Algorithm, p. 99.
- Yiannoudes, p. 50.
- Yiannoudes, p. 30.
- Yiannoudes, p. 30.
- Yiannoudes, p. 30.
- Yiannoudes, p. 31.
- Yiannoudes, p. 31.
- Alastair Parvin, ‘Architecture (and the Other 99%): Open-Source Architecture and the Design Commons’, Architectural Design: The Architecture of Transgression, 226, 2013, 90–95 (p. 95).
- Open Systems Lab, ‘The DfMA Housing Manual’, 2019 <https://docs.google.com/document/d/1OiLXP7QJ2h4wMbdmypQByAi_fso7zWjLSdg8Lf4KvaY/edit#> [accessed 25 April 2021].
- Open Systems Lab.
- Open Systems Lab.
- Carlo Ratti and Matthew Claudel, ‘Open Source Gets Physical: How Digital Collaboration Technologies Became Tangible’, in Open Source Architecture (London: Thames and Hudson, 2015).
- ‘An Introduction to WikiHouse Modelling’, dir. by James Hardiman, online film recording, YouTube, 5 June 2014, <https://www.youtube.com/watch?v=qB4rfM6krLc> [accessed 25 April 2021].
- Carlo Ratti and Matthew Claudel, ‘Building Harmonies: Toward a Choral Architect’, in Open Source Architecture (London: Thames and Hudson, 2015).
- Oliver David Krieg and Oliver Lang, ‘The Future of Wood: Parametric Building Platforms’, Wood Design & Building, 88 (2021), 41–44 (p. 44).
- Ratti and Claudel, ‘Building Harmonies: Toward a Choral Architect’.
- Carpo, The Second Digital Turn: Design Beyond Intelligence, p. 162.
- Robin Evans, ‘Towards “Anarchitecture”’, in Translations From Drawings to Building and Other Essays (从绘图到建筑物的翻译及其他文章), trans. by Liu Dongyang (Beijing: China Architecture & Building Press, 2018), p. 20.
- Walter Benjamin, Illuminations: Essays and Reflections (New York: Schocken Books, 2007), p. 12.
The following piece is transcribed from Fondamenta’s talk at the B-pro Open Seminar that took place at the Bartlett School of Architecture on the 8th December, 2021.
We are interested in the construction of spaces with a strong belief in research and experimentation, where building is the end to which architecture must strive to become itself, and technology is the tool used to reach this result. We question conventions and support contradictions; fascination for structure, and freedom from dogma are the premises of this research. Structure is the trace of space, it organises the program and generates the building. Governance through technology is the key to the creation of an architectural organism, we see our projects as opportunities to conduct research on structural systems and the use of materials. We push materials to and against their limits – we are into designing through a systematic approach, relative to structures, without forgetting that the ultimate user of this organism is the human being; we are glad to have seen four very interesting presentations. We connect with the work of Luigi Moretti a lot, who we deeply admire as an architect, being one of the first pioneers in understanding spaces as organisms, creating them with a scientific logic and having developed four precise categories to design them.
What is technology for us? It is an instrument that we face daily, we use technology to follow our purpose, and to reaffirm the central role of the Architect in the building process. Technology drives efficiency, precision and control through the entire process, allowing governance of the economy of the project. The central issue of the use of technology is always about WHO is responsible for the governance of it. We believe the answer is that the Architect should be able to take this role.
Today, we don’t want to talk about specific softwares and the use we make of them but rather point out the great opportunity that a specific use of technology could give Architects today. We were trained in a university founded on Vitruvian philosophy in which Architects must have a holistic approach to Architecture, being as much generalist as possible within the field of the discipline. Over time, we have witnessed a dismantling of the so-called “Generalist Architect”, in favour of over-specialisation in specific aspects of our discipline. The Architect has been relegated to a consultant, who concurs in order to create an architectural project. Instead, we believe the Architect must be the central figure, capable of managing the complexities of today’s world, through governance of many actors and aspects. This can only be possible, in our opinion, with the aid of technology. Our last resource is to believe a generalist Architect may still exist. . .
To achieve the latter, we use existing BIM (Building Information Modelling) technology to be superimposed with our customised system. For three years we have been testing a Vocabulary of codes and protocols that are applied to BIM and that become the common “language” inside the digital model that expresses the Architectural Project, which all involved actors have to learn and share. We, as Architects, are responsible for the governance of this centralised model and system, being the one creating the laws of the digitally-organised government. We didn’t start our practice directly with this idea, it was raised as a consequence of the first project we built and the impossibility we faced to have a central role in the process. Losing power and responsibility over the process with a negative impact on the projects was the consequence. We are still working on it daily to improve it, it is an ongoing process. If we have to depict with a diagram the shift between the approach we had at the beginning, and the approach we have now, this slide expresses it [indicates screen].
The centralised system we are looking for allows different actors to interact inside a given structure, with a given language crafted by us.
To get more into details, the above charts depict specific aspects of our customised Mother model. The strength of BIM is that it enables all consultants who are involved in the process to implement and add their knowledge and information inside a common, single-instance digital Model. Codes and rules were developed so as to share and communicate between the different disciplines, which belong to different worlds. The most important layer to be translated is that of economy. Each aspect of the project relates to an economical parameter that controls the cost of the projects. Starting from an existing software, we added our customised logic and vocabulary.
What we are seeing throughout our practice is that we can have control of the project from the very start. For the most part, BIM generally arises after an execution plan is in place. Instead, we deal with these premises from day zero – from concept phase – this is what makes enormous difference. Following this scheme, all actors begin to communicate at the very start, at the right time, without finding themselves in the position of compromise, but rather putting on the table all the topics that, if worked out at the right time, can surely bring the project to more radical expressions. Hence, there are incredible possibilities to push the projects to their limit, being able to build without it being jeopardised during an uncontrolled process.
We will show three different projects of ours. The first one, our first built project, is a winery in Piemonte (2018-2020).
In this project, our awareness of technology and its potential was limited and not yet evident. That is why we run this project without using BIM to solve design and governance issues. The winery project develops research on the pursuit of a seemingly impossible balance between different structural systems, which must coexist as one organism out of concrete and steel. We designed and optimised the shell system together with our engineer, making it work as structural truss to hold the concrete pitched roof while containing part of the program. The double steel formwork of the shells, poured in one single day without pause, was directly designed, drawn and sent to the manufacturer.
After this experience, we realised that we needed more technological support to be able to control the construction process in order to push forward more projects. Particularly dealing with aspects such as economics, time and money, but also sustainability of the process. This change of guard started with the series of projects we are building in Sicily, first among all 18018EH projects of houses near Noto. From this moment, we started governing the process with the aid of BIM – our instrument – from the beginning of conception.
This house is mostly underground, with only 30% of its surface exposed above ground meterage. We are trying to develop a three-dimensional project where the space develops in three axes, and all the load-bearing walls are made of local stone. The structural floor plan is created through a system of radius and circumference. Through the use of softwares, we were able to optimise the construction lines, turning them from splines to radius, working in accordance with the technical consultants to develop the BIM model. This is a snapshot showing the massive amount of information inside this model.
This is interesting because implementing information in a model is not enough to control it, there needs to be instrumental rules in order to make an architecture real. This project will be soon delivered to a construction company. Costs, money and time are essential points in our profession, in order to have the possibilities to realise our research, design cannot transcend from them. We are connected and interested in the economy of the project, which sustains architecture processes through awareness in governance and allows us to control our design according to cost.
It was incredible how we managed to control the project and design through our tools. For example, we like to show all these axonometric drawings – each code, of course, remains connected, with a clear Excel chart that reminds us of cost, quantities, and all the details that a specific part of the model has. Figuring out a way of communicating the mass of information that we were implementing in the digital model was another interesting aspect. This is something that we’re still developing to make it even more readable for the involved actors. Of course, there are just a couple of Excel spreadsheets connected to these axonometries!
In terms of design, we see the potential in technology as something that allows us to further push our research related to space and structure. For example, here, all the other walls will be made out of stone, blocks of stone that are one metre long, 50 centimetres high, and 30 centimetres in depth. For Grasshopper, we customised each one to come out with a sort of “abacus” of all the walls with specifications and a numbering system, then, delivered to a construction company.
This technology enables us to build within a certain amount of time. If we reflect on past projects, time is something that we really cannot negotiate – it is the hardest variable to negotiate today. Technology gives us the ability to control time more than any other aspect. We love to go back to the models, because we think that this “ping-pong” between the digital tool and the making process gives us an awareness of reality. We don’t have to lose control of what we are thinking and designing.
The last aspect that we are trying to show through this house – one of the projects already into construction since four months ago – is that we reached a certain level of governance of actors during the process from the beginning. This is a renovation, where we stripped out the existing building – the partition walls – but kept working with the existing concrete cage structure. We kept the load-bearing structure, made out of concrete, and we inserted a new steel structure, changing its form but keeping the volume untouched.
Wanting it to be a precise case study, we sat with our consultants and engineers from the very beginning. All the possible actors were involved from the embryonic phase and we designed together, trying to understand immediately all the potential realistic approaches that could be achieved.
I’ll just show a couple of snapshots of the model that we delivered to the construction company, pointing out that it is the same model we had from the beginning. From structures, H back, to installations, every element was designed with involved actors, long before the building process started on site.
It’s really important for us to underline that Architects have to be able to see and understand consultants and potential constraints as a possibility to further the design. For us, this was not something particularly easy to understand initially, because we were trained to see consultants and all other actors as part of architecture, and came in parallel to the project. Just like the scheme we showed, they are parallel lines that, at a certain point, intertwine. In that moment, you have a connection, and this connection has to be constant. Through this system we are developing, where each actor involved in the process has to be aware of the language we share in order to achieve the project.
This is just a snapshot of the house at the moment; we’ve stripped out the partition walls and it’s just the concrete.
To conclude, BIM has a deep social impact, giving back to architecture and architects the power they should have in the process. It is then up to us to create a social resistance and approaches to contemporary society.
The work of Leonardo and Laura Mosso provides a very early and original application of computation to architectural, urban, and territorial design. Although computers were actually utilised to develop their ideas (a rare event in the 1960s in Italy), the work possessed conceptual and political ambitions that exceeded both the simple (or even fetishistic) fascination for a new technology and the functional approach that conceives computers as tools to efficiently complete tasks. Rather, the computer was part of a proto-ecological approach in which artificial and natural elements worked together towards the emancipation of the individual and their environment. At the centre of their research was “Architettura Programmata”, defined as a “theory of structural design” dedicated to the design of elements, their connections, as well as a higher, meta-system which we could call “structure” in the sense that Structuralism defined this word. Computers were involved in this project under both a design and an ethical agenda to understand and define “ecocybernetic dynamic as a structure for a self-evolved language of the environment and of the form at various levels of complexity, inserted in an unforeseen chain of self-evolved cybernetics: from political cybernetic to cybernetic of information, as integrated instruments of evolution in a condition of direct articulated democracy”.
This paper will discuss how computational thinking and computers were employed in the work and research of Leonardo and Laura Mosso, by analysing three paradigmatic projects which tackled the notion of structural design at different scales and contexts. The first project will be Cittá Programmata (1967-70), a theoretical proposal for a new type of city. The project represents the first actual use of computers in the work of Leonardo and Laura Mosso. The second example will concentrate on a piece of research on Piedmont territory – the place in which they operated throughout their academic and professional careers. Although computers were not directly employed to carry this research out, the approach to territorial analysis and planning employs a form of algorithmic thinking which impacts on both how the territory is read and how it could be re-imagined. Finally, the proposal for the restoration of block S.Ottavio in the historical centre of Turin shows a very innovative use of computers to intervene on historical artefacts of relevant cultural value, as well as the possibility to use computers to manage the future life of a building.
Structuralism played an important part in the work of Laura and Leonardo Mosso, and it is an essential element in understanding their conceptualisation of structures and the role that design and computation had in it. A slightly left-field but very fruitful interpretation of Structuralism was produced by Gilles Deleuze in 1967, at the time Leonardo and Laura were intensifying their interest in computers. Deleuze emphasised the role of emptiness, more precisely, of the “zero” sign as a mechanism for the transformation and articulation of structures. The notion of empty structure and zero offer a dynamic interpretation of Structuralism that is not only relevant to computational thinking, but can also clarify how the structures designed by Mosso can be understood as dynamic and adaptive.
Early Experiments with Computers in 1960s Italy
Before delving into the actual discussion, it will be useful to quickly sketch out some of the cultural trends operating in Italy in the 1960s to better contextualise how Leonardo and Laura Mosso arrived at their “Architettura Programmata”.
“Architettura Programmata” directly refers to the exhibition “Arte programmata. Arte cinetica. Opere moltiplicate. Opera aperta” organised by Olivetti in 1962. The show was curated by Bruno Munari and Giorgio Soavi, with an accompanying catalogue edited by Umberto Eco. It displayed works by a series of artists, including Enzo Mari, who generated art procedurally, opening up a different mode of production and reception of works of art, also inspired by Eco’s Open Work. In the same period, Nanni Balestrini was also experimenting with computers to generate poems. These two examples are perhaps useful in helping to focus on some lesser-known aspects of Italian post-war culture, which is often mentioned for the work in cinema, architecture, art, but rarely for computation or scientific work in general. Along these lines, it is also worth mentioning the cybernetic group operating in Naples under the guidance of Prof. Eduardo Renato Caianiello, who maintained regular contact with MIT and Norbert Wiener. It is in this more international and open environment that we should position the research of Leonardo and Laura Mosso.
Leonardo studied architecture in Turin, a very active city that led the Italian post-war economic boom, thanks to the presence of Fiat, the car manufacturer and one of the largest Italian factories. After graduating, Leonardo won a scholarship to study in Finland where, eventually, he started working in Alvar Aalto’s studio around 1958. From then on, he became the point of reference for most of the works that Aalto designed for Italy—such as the design of a residence for the Agnelli family (the owners of Fiat) and the Ferrero factory. A more international profile also characterised the figure of Guiseppe Ciribini, with whom both Laura and Leonardo also collaborated. Ciribini concentrated on the modernisation of the construction industry, focusing on prefabrication and modular design. His work was not limited to Italy and expanded to a European scale through his involvement with the European Coal and Steel Community (ECSC, or CECA in Italian, the precursor of the European Union) to devise international standards for prefabrication. Leonardo and Laura Mosso also established connections with Konrad Wachsmann, incidentally Giuseppe Ciribini’s predecessor at the Ulm School of Design, invited by Tomas Maldonado in 1958. Finally, Leonardo and Laura Mosso were also involved in the early experiments with computer art (which had developed in Croatia since the early 1960s) through the magazine New Tendencies.
In all these experiences, computation played an increasingly central role. In the case of Ballestrini or for the scientific research developed in Naples, computers were actually utilised, but in other cases, the work only consisted in speculation over what tasks and possibilities could be performed and unleashed. Leonardo and Laura Mosso are among the small group of architects and artists who did make use of computers in their work. With the help of Piero Sergio Rossatto and Arcangelo Compostella, two projects utilised computers to simulate and manage their transformations. Throughout almost two decades of using computers in their work, Leonardo and Laura Mosso developed an approach that was never guided by technocratic notions of efficiency. Rather, the philosophical implications of computing architecture and the political role that information and computation could have brought to a project and society in general constituted their main interest in this new technology. The computer as used in the Mossos’ work was in fact at the service of larger cultural project that aimed at distributing, rather than concentrating, power. Computers were an instrument for change, whereas the values of efficiency and sheer industrialisation appeared to be ways to fundamentally preserve the status quo, by simply making it run more smoothly. Rather than improving how architecture could better fulfil its role under the tenets of a capitalist, industrialised economy, Leonardo and Laura wanted to change the rules of the game itself; the computer, therefore, had to play an almost moral role in radically overturning the mechanisms regulating architecture and its use.
Central to their research was the close relationship between philosophical ideas (Structuralism), design language (which particularly concentrated on discrete elements connected through reconfigurable, dynamic nodes), and computation. Leonardo and Laura Mosso’s approach to Structuralism was already open to dynamic, cybernetic influences and, for this reason, it may be interesting to read it against the famous writing that Gilles Deleuze dedicated to the same philosophical movement.
The Dynamics of Structural Form
Culturally, the post-war years were characterised by the diffusion, particularly in Italy and France, of Structuralism; generally understood as a philosophy of structures rather than functions. Structures could be organised in more general systems – of which natural language represented the most complex, paradigmatic example. Linguistics was indeed the domain of Structuralism, and the source from which most of its fundamental ideas were derived. From Saussure’s Course – indicated as the first structuralist text – to Barthes, Eco, Levi-Strauss, the Bourbaki group, Althusser, and also Foucault and Lacan, structuralist thinking extended beyond the linguistic domain to provide a framework to re-conceptualise other disciplines such as anthropology, psychoanalysis, mathematics, history, or politics.
Broadly speaking, the definition of a structure consisted of two steps: the determination of its constituent parts (taxonomy) and the definition of the mechanisms that would govern the relations between parts and their transformation (grammar). Critics of Structuralism often rebuked this particular approach to structures for its excessive formalisation and the strictness of its deductive logic. Such criticism tended to depict structuralism as a mechanical, overly linear theory of systems, resulting from the perhaps excessive importance attributed to linguistics. Perhaps such characterisation of structuralism paid too little attention to the more transformative aspects of the theory: the dynamics of change and transformation. These are present in all the major structuralist thinkers; however, Gilles Deleuze provided an original overview that concentrated on the open, topological, and playful aspects of structures which is useful to briefly summarise here. In Deleuze’s “How Do We Recognize Structuralism?”, originally written in 1967, Structuralism was detectable through six different criteria: symbol, local/positional, differential/singular, differentiation/differentiator, serial, and the empty square. Throughout the analysis, the emphasis is on transformation rather than permanence, on the mechanisms that guarantee a structure can operate by straddling between the real and the imaginary in order to transform reality and be transformed by it.
We will return to Deleuze, particularly his understanding of the notion of “zero” which offers an interesting frame in which to conceptualise the role that structures played in the work of Leonardo and Laura Mosso – and, particularly, how physical construction nodes were instrumentalised to attain a structural language able to change and be appropriated (or “spoken”) by its users. Before dwelling further on this aspect of their work, it is important to point out that the work of Jean Piaget – an author often quoted in Leonardo’s and Laura’s writings – also offered a dynamic reading of structures and Structuralism in general. Laura and Leonardo often made use of Piaget’s characterisation of structures being composed of three main characteristics: wholeness, transformation, and self-regulation. In Piaget’s work, we also we find an open, interactive, “proto-cybernetic” reading of Structuralism marked by a relational understanding of the connections between environment, cognition, and symbols. Particularly, the notion of assimilation outlined by Piaget in The Construction of Reality in the Child outlined a cognitive model based on continuous feedback between reality and the child’s development – an image that brought Structuralism much closer to cybernetics. An eco-cybernetic approach to planning was often also advocated by Laura and Leonardo. These initial definitions are helpful, not only in framing the work of the Mossos in relation to the cultural milieu in which they operated, but also in understanding how computation was conceptualised in their projects to translate notions of structure, node, transformation.
As mentioned, Deleuze’s survey offers a particular vantage point to understand how Structuralism dealt with change and transformation, and how this can help to frame the role that structures and nodes have in the research of Laura and Leonardo Mosso. Deleuze dedicates particular attention to the notion of the “zero” sign in Structuralism; the “zero” sign is understood as an empty place in structure, determined positionally rather than semantically, that allows transformations to occur. The empty place in a structure guarantees the possibility of its transformation in a way which is analogous to the role of empty squares on a chess board. The structure is understood as a symbolic object. Symbols are here understood according to the definition provided by C.S. Pierce’s semiotics; that is, structures have an arbitrary character that does not attempt to find the essence of the object of investigation, but rather to construct it. In Deleuze’s words: “[the structure does not have] anything to do with an essence: it is more a combinatory formula [une combinatoire] supporting formal elements which by themselves have neither form, nor signification, nor representation, nor content, nor given empirical reality, nor hypothetical functional model, nor intelligibility behind appearances”. The structure is always a third, encompassing element, beyond the real and the imaginary, that allows the structure “to circulate”. In other words, the elements of a structure can only be determined relationally, as “[they] have neither extrinsic designation, nor intrinsic signification”. As the order of the structure is more important that its meaning, not only is the space (or spatium, as Deleuze refers to it) a central medium for the articulation of relations and transformations, but is best described topologically, in the sense that the function of such a spatium is to logically order elements so that specific, empirical objects can occupy the different squares of the structure. The final element to note in Deleuze’s analysis is the “wholly paradoxical object or element”, that is, the connective element that allows different structures or series to communicate with and orient each other in order to perform on different levels, beyond the purely symbolic one. Such an element is defined by Deleuze as the “object = x”; the “zero” sign par excellence; the “eminently symbolic” object that injects dynamic qualities into structures and therefore allows them to work.
Leonardo and Laura Mosso dedicated large parts of their architectural research to the roles that connecting elements, or nodes, had in articulating structures. Such research produced four different types of nodes which informed their work and can be seen at work in the three projects discussed in the second part of this paper. Deleuze’s consideration on structures help us frame the Mossos’ research as well. The node in a structure is the element that allows transformations to occur. Pieces can be detached, substituted, or removed according to the possibilities and constraints set by the node connecting them. There is therefore an analogy between the physical nodes of a structure and the mechanisms of transformations at work in the philosophical concept of structure. Borrowing from Deleuze’s description, the physical node becomes the “object = x”, the “zero” sign; that is, not simply the element that makes change possible, but also the element that is syntactically operative and open in order for meaning to emerge. The analogy between the two manifestations – physical and philosophical – structure is poignant to grasp the Mosso’s work: nodes are often literally “zero” signs, voids as in the case of the particular type of node developed for Cittá Programmata is literally organised around a void, an empty space. By straddling between its physical appearance and its philosophical interpretation, the node acts structurally, that is, beyond its purely empirical presence, the node is a device that orders physical elements logically. . In both accounts of structures, the minimal unit is the phoneme – “the smallest linguistic unit capable of differentiating two words of diverse meaning” – which Leonardo and Laura put at the centre of their approach to structures by speaking of “phonetic” and “programmed structures”. This approach was already visible in the first example of “programmed architecture”, the Chapel for the Mass of the Artist in Turin (1961-63) in which a static node connected together 5cm x 5cm wooden studs to produce a highly varied pattern for the interior of the Chapel. In successive projects, nodes quickly grew in complexity in order to achieve more articulate and varied configurations, as well as allowing the end user and community to be able to adapt them for future uses. Such an architectural agenda demanded a new type of node that began to be articulated as a void, an “empty square” so to speak, around which the various elements aggregate (fig.). The morphology of this new type of node consisted of a virtual cube – a void – whose eight vertexes could be reconfigured around smaller voids, each able to link together four members. None of the members physically intersected (making the implementation of changes easier) and were organised around a series of voids of different sizes. These physical and conceptual voids held some analogies with the “object = x” Deleuze spoke of in regards to Structuralism; the final configuration was dynamic, a sort of system to let the structure circulate, to make transformation possible. In other words, such an approach to structure transformed the spatial model of representations from a strictly geometrical system to a topological one in which relations between objects took precedent over the presupposed semantic qualities.
It is also along these lines that we can read the introduction of computation into the work of Leonardo and Laura Mosso. The computer became the perfect instrument to both manage the structural logic of the design and give it the political agency the two architects had been seeking through their notion of programmed architecture. The next section will analyse three paradigmatic projects in which the conceptual issues highlighted can be seen at work.
Cittá Programmata, 1967-70.
Cittá Programmata is one of the most iconic projects developed by Leonardo and Laura Mosso, a manifesto that encapsulates some of the key aspects of their work; that is, the potential for a structural approach to design to provide an environment for social and political self-determination. To implement their agenda of political and spatial self-determination, Leonardo and Laura introduced the computer, which represents the other radical aspect of this project. The computer played both an operational and a moral role in enabling the appropriation and transformation of the users’ habitat. Strictly speaking, the project consisted of a series of physical models and computer-generated drawings for an entire city and its possible transformations. The city was structured through a series of cubical modules (or “voxels”) of 6m x 6m x 0.5m that could co-evolve with the life of the city and its inhabitants, resulting (as the models and drawings showed) in an interrupted field of variously extruded elements, each composed by structural elements variously transformed.
The research for Cittá Programmata took place in a rich cultural environment in which the work of Laura and Leonardo stood out for its original take on some of the topics that animated the architectural debate of the time. As mentioned, the post-war Italian scene was characterised by a growing importance of Structuralism in all aspects of culture. On the one hand, Structuralism guided the introduction of linguistics and semiotics as a general field of study, as well as their application to architectural and urban analysis. This line of inquiry sought to detect the underlying principles of architectural form in itself and in its relation with its context. At the other hand of the spectrum, a more pragmatic understanding of structural thinking was animating the debate on pre-fabrication and modular design, to renew the construction industry and fulfil the demand to modernise the Italian landscape. It is between these two main interpretations of the notion of structure in architecture that Cittá Programmata can be understood, as it proposes a different conception of language and structures.
Leonardo and Laura Mosso saw in the semiotic approach to architecture an excessive interest in meaning, both in its relation to the internal history of architecture and in context. Against the backdrop of semantic studies on architecture, Cittá Programmata proposed a more structural approach to language and its formalisation; a “phonological” system that would enable its users to ‘speak’ their collective mind through the groups of structures the architects provided. Pre-fabrication, on the other hand, was indeed a rich field of investigation – as mentioned, Leonardo and Laura Mosso had been in close contact with Giuseppe Ciribini. However, prefabrication was committed to a model of society that privileged economic values (through the minimisation of costs, for instance) over political, cultural and social ones. Indirectly, their critique of pre-fabrication was also a critique of the notion of programme (“programma edilizio”), understood as an excessively functional approach to design. The brief – the document through which to implement a building programme – fixed the use of structures or, at best, described a limited number of activities that a piece of architecture could house over a limited period of time. The formalisation of such an approach to programme usually resulted in a neutral outcome which favoured the design of a generic spatial container which, in principle, could adapt to future needs. Leonardo and Laura critiqued this view of design both on the basis of the vagueness of the mechanisms for programmatic determination (future activities may be impossible to predict in advance) as well as for the generic architectural response. In opposition to it they proposed a structural approach that did offer implementable choices (as opposed to programmatic vagueness) and therefore was not limited to regulating quantitative growth, but could also take into account the qualitative aspects of spatial structures. Finally, programme was also critiqued from a political point of view, as it was identified as the political instrument that guaranteed an asymmetrical distribution of power between users and designers.
Cittá Programmata imagined an environment in which the relation between users and architects was not hierarchically organised, but rather more radically and horizontally distributed. Here, both the programmatic and semantic critique that animated the Mossos’ approach converged. The aim to generate an environment based on a horizontal distribution of power called into question the role that semiotics could play in designing structures. The analogy proposed is once again with language. As for immaterial notions, language and architecture (understood as body of knowledge) are inherently public, they exceed anyone’s ability to claim ownership of them or control them. Both the linguist and the architect can only play with the systems of signs constituting their disciplines in order to make them public and accessible. Contrary to the semiotic studies of architecture which concentrate on the internal mechanisms and references of architectural language, Leonardo and Laura Mosso proposed a rather more “extroverted” approach interested in opening architecture up and inviting users to participate in the creation of their own environment. The architect was “at the service” of architecture, rather than a custodian of the arcane mechanisms of architectural language. In a way, we can say that the position taken was reminiscent of Saussure’s distinction between langue and parole: whereas semiotic studies in architecture appear to privilege the importance of the langue, in Cittá Programmata, Leonardo and Laura Mosso worked to maintain a dynamic relation between the two terms of Saussurean categorisation:
Architecture, understood in a traditional sense, cannot be a language; that is, it cannot speak by itself. Similarly, we cannot say that the work of a linguist on language is a language … Architecture is at [the] service of language … in the same sense that a language services the community of speakers when it is spoken; that is, when architecture becomes “a system of transformations” or possibilities, from which it is possible to generate infinite messages.
Mosso and Mosso
It is in this context that the computer was introduced, both to support the management of the city and to simulate its future configurations. The actual machine utilised was a Univac 1108 owned by the Politecnico of Milan and programmed by Piero Sergio Rossatto – an engineer and programmer at Olivetti – with Arcangelo Compostella. The stunning drawings generated by the Univac (now part of the Centre Pompidou’s permanent collection) showed the possible growth patterns generated from an arbitrary string of signs placed at the centre of the drawing. Two parallel lines of pre-allocated units (*) and voids (-) constituted the starting input for the simulation, which could either proceed in a sequential growth, on the basis of a probabilistic algorithm, (fig.XX) or randomly (fig.XX). The process of algorithmic growth did not take place in a vacuum, rather constraints could be programmed in making growth sensitive to contextual information.
Landscape, Structure and History (1980-1986)
A second type of node Leonardo and Laura Mosso had been working on were a kinetic, self-managed, and elastic universal structures(Strutture autogestibili e complessizzabili a giunto universale elastico). Since the beginning of the 1970s, as part of their research on the use of different types of nodes to articulate transformations in physical structures, they had been testing this particular type of node at different scales and in contexts. The research started with the academic work that Leonardo carried out with his students at the Politecnico in Turin, then through commissons such the “Red Cloud” (Nuvola Rossa), an installation completed in Carignano Palace in 1975 in which these nodes found one of their most convincing and poetically powerful applications. This large piece consisted of a complex structure made up of individual elements connected through elastic joints, which allowed the architects to build an undulating mesh suspended between the visitors and the frescos of the palace. These elastic structures were tested at different scales: for instance, between the end of the 1970s and the beginning of the 1980s, Laura and Leonardo would put their kinetic quality to the test by using them as props accompanying the movement of the bodies of contemporary dancers, both in their work with the Conservatorio G. Verdi in Turin (1978) and the performance staged in Martina Franca (1980). It is, however, the territorial scale which is of particular interest in this discussion, since it highlights an original understanding of how structures can perform algorithmically and because of the unusually large scale of this research.
Here, particular reference is made to the research carried out between 1980 and 1981 under the broad agenda of “methodological work aiming at devising a system of signs to program both at the level of the territory and the city”. The results of this methodological analysis of territorial structures would also inform a subsequent research project and exhibition titled “Landscape, Structure, and History”, which tested their structural approach to territory on the local landscape of Piedmont, its rural cultures, and their relation with their surroundings, with a view to devising a strategy for preservation. Perhaps it might appear unusual for avant-garde architects to dedicate their research to the rural, historically-layered territory of Piedmont. On the contrary, local forward-thinking architects and engineers had already focused on vernacular architectural expressions in the local countryside: Carlo Mollino extensively studied and recorded examples of Alpine vernacular architecture in Valle D’Aosta, and Giuseppe Ciribini – whose work on the industrialisation of construction has already been mentioned – also paid attention to the spontaneous architecture of Alpine and pre-Alpine territories. Some of these interests in rural and vernacular architecture were gathered together by another Torinese architect, Giuseppe Pagano, in his famous exhibition “Continuity – Modernity”, in 1936, for the 6th Triennale in Milan.
The Mossos’ research on territorial structures consisted of both drawings and physical models of specific areas of Piedmont (Canavese and Carignanese). The work mapped and recorded the landscape of Piedmont by positioning a series of kinetic structures over a map of the existing territory. The structures consisted of a series of elements connected through elastic, kinetic nodes that allowed each element complete freedom of rotation around each vertex. The final configuration of each structure emerged from the mediation between their internal properties (length of the elements, arrangement, type of nodes) and the cartographic representation of the landscape. The drawings took this relationship to more radical conclusions: the landscape was further abstracted and re-coded through a structural approach which adapted to different contexts. Rather than an image of a superstructure, the re-codification of the landscape through models and drawings struck a complex balance between the algorithmic approach and the context.
In this particular project, structures are understood as organisation principles rather than physical constructions. Earlier, we spoke of a algorithmic use of structural thinking, a quick definition that requires unpacking. An algorithm is a set of instructions that, once applied to a set of input data, will perform a finite number of operations to return an output. Regardless of the complexity of the operations performed, an algorithm recodes the input data into a new set of data. Chomsky’s generative grammar, for instance, could be seen as a recursive (continuous) series of algorithms that rewrites any given statement of a natural language to produce new linguistic statements. The superimposition of Laura’s and Leornardo’s structures on a map of Piedmont countryside operated in a similar fashion and, therefore, could be interpreted as an algorithmic recoding of the territory. The input data was constituted by the information recorded in the cartographic representations of the landscape, whereas the kinetic structures acted as analogue algorithms that recoded the input data according to the vast (yet finite) number of configurations allowed by their physical characteristics (length and number of members, type of joints). In short, the physical structures deployed rewrote the landscape according to a precise set of rules; more poetically, we can say that the elastic node structure allowed the landscape to speak in the language of the structures superimposed onto it; an image that Laura Mosso also evoked when she wrote about developing methods to “make the structures whistle”.
Contrary to stricter interpretations of Structuralism, the type of algorithmic approach proposed here was not merely deduced from internal, formal rules (that is, the physical constraints set by the elastic nodes); rather it emerged from a more iterative, open relationship with the context (abstracted through cartographic representations). The results of the process set up were particularly legible in the physical models: the kinetic structures made up of interconnected springs were laid out on the map to return a ‘structural re-reading’ of the landscape. A new, structural image of the territory emerged from the interaction between nodes and territory.
The research on territories that Laura and Leonardo Mosso completed allows us to make a series of considerations on these algorithmic operations, their formal qualities, and the implications they give rise to. First, through a structural, algorithmic approach to territory, the research rejects distinctions between natural and artificial in favour of a more holistic approach to landscape – and yet, one describable through a set of finite operations. The constraints embodied in the physical structures do not decisively distinguish between artificial and natural, symbolic and productive, and thus support Leonardo and Laura Mosso’s call for the kind of expanded notion of ecology they had been advocating for, both in projects and publications (through, for instance, the publication titled La Nuova Ecologia). The structure is the symbolic device that catalogues and organises the whole of the territory (here understood as superseding dichotomies such as urban/rural, artificial/natural), establishing principles for its preservation and transformation Similarly, algorithmic re-writing provides a diachronic reading of the territory that is re-organised along structural rather than chronological vectors. The different nodes of the elastic structures are positioned on the map to establish connections between artifacts built in different times in order to give rise to new relations between them. Finally, there is the function performed by the elastic structures as analogue algorithms. We have already seen how an algorithm can be understood as a form of rewriting and transformation of an existing condition (input data). The types of operations performed by an algorithm are always precise (determined by the rules programmed in the algorithm), executed in their entirety (the algorithm goes through all the steps scripted to return an output), and yet partial, as the algorithm can only survey a dataset according to the set of rules that form the algorithm itself. The constraints inbuilt in the elastic kinetic nodes allow them to only perform a vast, but finite set of movements; that is, only a subset of all the signs contained in the maps of Piedmont can be computed by the physical structures-algorithms. In short, an algorithm generates a specific representation of the object it is applied to.
To better grasp this last point, we can draw an analogy between real objects (such as buildings) and their orthographic representation. For instance, a section through a building can only return a partial image of the object it investigated, and yet how a section is drawn follows precise and rigorous rules that determine what and how the building will be captured in the section. But the section is a sign-object, not a building; it elicits further manipulations by either applying different sets of criteria (e.g., by concentrating on the structural, programmatic, material qualities of the building) or by changing the very parameters that generated it (changing the position of the section plane or the conventions applied). The approach developed to the Piedmont territory by Leonardo and Laura Mosso makes aspects of this landscape intelligible through the production of new signs which, in turn, make it amenable to further manipulations. It is important to notice that all operations performed by Laura and Leonardo are performed on a cartographic representation of the territory; photographs and other cultural aspects of the areas such as the name of places are complementary, rather than primary, information. Cartography is itself a coded, notational (rather than mimetic) representation of the territory. As a medium it therefore lends itself to the operations of re-coding and re-writing, since it is already a semiotic system; on the other hand, it acts as a recipient of the new codification of the landscape generated through a structural reading.
Finally, the structure-algorithm becomes a marker of change, as the instrument through which modifications, and, in general, any metamorphic transformation of the territory can be foregrounded, read, and made tractable in order to preserve it or alter it. The research developed by Laura and Leornardo Mosso shows that a structural approach through algorithmic thinking should not only be confined to new, pristine domains, but can also offer innovative ways to interpret and intervene in historical contexts. The last project discussed – the proposal for S. Ottavio block in the historical centre of Turin – will further reinforce this point.
S. Ottavio Block, Turin, 1980
The commission for a study of the block located in the historic centre of Turin was received in 1978 and became an important, yet entirely forgotten chapter in the story of both Leonardo and Laura Mosso’s production and the integration of digital technologies in architecture. On the one hand, the brief for the project was a rather common one for Italian architects, whose practice often confronted (and still confronts) historical artefacts. Leonardo and Laura, however, saw in this commission an opportunity to advance their research on structures as well as on the use of computational tools. For purposes of simplicity, we can artificially divide the project between the physical proposed interventions and the immaterial, data-driven ones.
The physical restoration of the block consisted in a series of more traditional interventions to reinforce the old brick walls, as well as the insertion of new levels to convert the existing spaces into inhabitable housing units. The new structures in steel and wood were elegantly laid out at a 45-degree angle, to mark a clear distinction between pre-existing and new elements. The type of node deployed in this instance was also a dynamic one, however, the only permissible movement was to slide along one of the orthogonal directions of the structure. Though the dynamics of nodes were limited (in comparison to the conceptual experiments at territorial scale), they allowed users to alter and self-organise their habitat. By deploying the same type of node at different scales and through different materials (aluminium, wood, and plexiglass), users could appropriate the environment both at the architectural and interior scale.
Perhaps the most radical proposal of this research was the organisation of the conceptual side of the project. A computerised system was going to be set up to monitor and maintain the block. A proto-digital twin, the system would map all the elements of the project and generate a database in order for both individual users and the municipality to control, repair and maintain the whole block. For the programming of the whole system, Piero Sergio Rossatto – who had worked with Laura and Leonardo for the Citta Programmata – was consulted. The spatial representation of the block in the digital model followed the logic of voxels: a three-dimensional grid of individual cubes that provided a system of coordinates to locate every element of the project, existing or proposed, architectural or infrastructural. In Rossatto’s scheme, the project would be surveyed starting from the ground level (z=0 in the digital model) and gradually moving towards the roof by increasing the z-value in the voxel grid. Every intersection between the voxel grid and an element of the project would be recorded.
Although the project was not well received by the local administration that could not fully grasp the innovative approach, eventually shying away from a unique opportunity to radically rethink the relation between digital technologies and historical artefacts, the project illustrated a different, complementary fact of Leonardo and Laura Mosso’s approach to algorithmic form.
As mentioned, the project applied digital technologies to pre-existing architectural artefacts protected by preservation laws. Whereas digital technologies are invariably understood as the instrument to deliver the “new” or the “radically different”, or even to make a tabula rasa of pre-existing notions, this project showed a more nuanced, and yet still radical side of digital technologies, which could coexist with and complement the delicate pattern of a historical city.
The structural approach, which continuously developed throughout several decades of research, here resulted in an abstract grid – a field of voxels, to be precise – that acted as a monitoring system allowing users to appropriate and control their own habitat. In the course of their research, Leonardo and Laura developed a physical model of the virtual voxel field that did not include any of the physical structures designed. The model possessed a very strong sculptural quality, but, most importantly, also showed the power of the algorithmic approach they had developed. On the one hand (and similarly to the experiments carried out in coding the Piedmont territory), the logic of the structure not only enabled its own transformation, but also determined its aesthetic qualities. The algorithmic logic guiding its own re-writing (in this case represented by the rhythm of the voxel field) returned a new type of form; an algorithmic form. As the model clearly showed, the logic of the voxel field implied a space without discontinuities or interruptions; saturated with data, the model was “all full” (as Andrea Branzi would have it), a solid block of data. As such, the research and proposal for the S.Ottavio block represents one of the earliest attempts to think of design straddling between physical and digital environments – a concept that could only be implemented through a structural approach to design whose robustness would allow it to extend to immaterial representations of space.
The work of Leonardo and Laura Mosso not only constitutes an excellent example of very early work with computers in architecture, but also provides a rich framework through which to problematise the issue of algorithmic form. The close relationship between design, philosophy, technology and politics not only forms a complex and rich agenda, but also expands the use of computers in design well beyond a functional focus on increasing efficiency and profits. Perhaps, this is one of the aspects of their work that still resonates with contemporary research on algorithmic design: the complex relationship between ideas and techniques, and the use of computation as an instrument for change. Computation was more than a vehicle to implement their radical design agenda, it was also tasked with implementing specific ethical values by orchestrating the interaction between architects, users, and built environment. In many ways, computation, and the algorithmic forms it engendered, was utilised by the Mossos to perform one of its original and most enduring tasks: to logically order things and, therefore, to conjure up an image of a future society.
In memory of Leonardo Mosso 1926-2020.
 L. Mosso & L. Mosso, (1972). “Self-generation of form and the new ecology”. In Ekistics – Urban Design: The people’s use of urban space, vol.34, no.204, pp.316-322.
 Deleuze’s text on Structuralism, however, was only published in 1971, so the connection between the two architects and the French philosopher is coincidental.
 U. Eco, The Open Work, Translated by A. Cancogni. 1st Italian edition published in 1962. (Cambridge, Mass: Harvard University Press, 1989).
 R. Bottazzi, Digital Architecture Beyond Computers: Fragments of a Cultural History of Computational Design (London: Bloomsbury Visuals, 2018).
 L. Mosso & L. Mosso, “Computers and Human Research: Programming and self-Management of Form”, A Little-Known Story about a Movement, a Magazine, and the Computer’s Arrival in Art: New Tendencies and Bit International 1961-1973, edited by M. Rosen. (Karlsruhe, Germany: ZKM/Center for Art and Media; Cambridge, MA: MIT Press, 2011) 427-431.
 G. Deleuze, “How Do We Recognize Structuralism?”, Desert Islands and Other Texts 1953-1974, Ed. D. Lapoujade, transl. by M. Taormina. (Los Angeles, CA: Semiotexte, 2004). Originally published in F. Chatelet (ed.) Histoire de la philosophie vol. VIII: Le XXe Siècle. (Prasi: Hachette, 1972), 299-335.
 J. Piaget, Structuralism. Translated and edited by C. Maschler. (London: Routledge and Kegan, 1971, 1st edition 1968).
 E. Von Glaserfeld, “The Cybernetic Insights of Jean Piaget”, Cybernetics & Systems, 30, 2 (1999) 105-112.
 J. Piaget, The Construction of Reality in the Child (New York: Basic Books, 1954; 1st Edition Neuchâtel, Switzerland: Delachaux et Nestlé, 1937)
 G. Deleuze, “How Do We Recognize Structuralism?”, Desert Islands and Other Texts 1953-1974, Ed. D. Lapoujade, transl. by M. Taormina. (Los Angeles, CA: Semiotexte, 2004). Originally published in F. Chatelet (ed.) Histoire de la philosophie vol. VIII: Le XXe Siècle. (Prasi: Hachette, 1972), 173
 Ibid., 173
 Ibid., 184
 Ibid., 176
 L. Mosso & L. Mosso, “Architettura Programmata e Linguaggio”, La Sfida Elettronica: realtá e prospettive dell’uso del computer in architettura (Bologna: Fiere di Bologna, 1969) 130-137.
 L. Baccaglioni, E. Del Canto & L. Mosso, Leonardo Mosso, architettura e pensiero logico. Catalogue to the exhibition held at Casa del Mantegna, Mantua (1981).
 L. Castagno & L. Mosso, ed. Paesaggio, struttura e storia: itinerari dell’architettura e del paesaggio nei centri storici della Provincia di Torino Canavese e Carignanese. (Turin: Provincia di Torino Assesorato all Cultura, Turismo e Sport, 1986).
I was asked by Mollie Claypool to curate the second issue of Prospectives Journal as an ideal follow up to leading Research Cluster 0 at B-Pro in the academic year 2020/21. As such, this issue is a collection of positions that respond to my research interest during that year.
In fact, my initial objective with RC0 was to research ways of applying computational tools to housing design for high-rise typologies: the aim was to update modernist housing standardisation derived from well-established rationalist design methodologies based on statistical reduction (such as in the work of Alexander Klein and Ernst Neufert), with the computational tools available to us now.
While the outcomes of this research were indeed interesting I was left with a sense of dissatisfaction, because it was very difficult to achieve architectural quality using purely computational tools – in a sense I felt that this attempt at upgrading modernist standardisation via computation didn’t guarantee better quality results per se, beyond merely complexifying housing typology and offering a wider variety of spatial configurations.
In an essay I published in 2019 (which in many ways inspired the curation of this Journal), I declared my interest to be in the use of computational tools not for the sake of complexity – formal or programmatic – but for increasing architectural quality, while decrying that the positions expressed by the so-called first and second digital revolutions, at the level of aesthetics at least, seemed too invested in their own self-proclaimed novelty. My interest was in rooting them in a historical continuum, with established architectural methodologies; seeing computational design as an evolution of rationalism.
This is why I wanted this journal to be about architectural form, and not about technical aspects of computational design: there is an urgent need to discuss design traditions connected to computational design, as an inquiry on “best practices” – that is, historical cases of what an algorithmic form has been and can be.
Any discussion on architecture implies a twin focus, on the one hand, on the technical aspects of construction and the tools of design, and on the other, on how these are interpreted and sublimated by the artistic sensibility of an author. Ultimately, what’s interesting about architecture as the discipline of constructing the human habitat is how it is capable of producing a beautiful outcome; and in architecture, perhaps more than any other practice, the definition of beauty is collective. To be able to establish what’s beautiful, we need to develop common hermeneutic tools, which – much like in art – must be rooted in history.
In light of this, I’m delighted with the contributions to this Journal, which offer a concise array of historical and contemporary positions that can help construct such tools. Many of the essays presented here offer a much needed insight into overlooked pioneers of algorithmic form, while others help us root contemporary positions in an historical framework – thus doing that work necessary for any serious discipline, technical or artistic, of weaving the present with the past.
My hope is that those individuals or academic institutions who are interested in how we can use emerging computational tools for architecture can re-centre their work not just on tooling and technical research but on architectural form, as the result of good old composition and proportion. The time is ripe, in my view, for bridging the gap between computational fundamentalists who believe in the primacy of code, and those with more conservative positions who foreground good form as the result of the intuition and inclination of a human author, remembering that an architectural form is only interesting if it advances the quality of life of its inhabitants and continues to evolve our collective definitions of beauty.
What’s the Hook? Social Architecture?
Isa Genzken’s work can be seen as a synthesis of the “social” and the “object” – a visual-sculptural art that reflects on the relationship between social happenings and the scale of architectural space. She was also one of the early explorers in the use of computation for art, collaborating with scientists in the generation of algorithmic forms in the 70s. But what is the social object? What can it mean for architecture? Just as Alessandro Bava, in his “Computational Tendencies”, challenged the field to look at the rhythm of architecture and the sensibility of computation, Roberto Bottazzi’s “Digital Architecture Beyond Computers” gave us a signpost: the urgency is no longer about how architectural space can be digitised, but ways in which the digital space can be architecturised. Perhaps this is a good moment for us to learn from art; in how it engages itself with the many manifestations of science, while maintaining its disciplinary structural integrity.
Within the discipline of architecture, there is an increasing amount of research that emphasises social parameters, from the use of big data in algorithmic social sciences to agent-based parametric semiology in form-finding.  The ever-mounting proposals that promise to apply neural networks and other algorithms to [insert promising architectural / urban problem here] is evidence of a pressure for social change, but also of the urge to make full use of the readily available technologies at hand. An algorithm is “a process or set of rules to be followed in calculations or other problem-solving operations, especially by a computer”. It is a finite sequence, well-defined, with performance based on the length of code – how fast and best can we describe the most. In 1975, Gregory Chaitin’s formulation of Algorithmic Information Theory (AIT) reveals that the algorithmic form is not anymore what can be visualised on the front-end, but “the relationship between computation and information of computably generated objects, such as strings or any other data structure”. In this respect, what stands at the convergence of computable form and the science of space is the algorithmic social object.
Social science is the broad umbrella that encompasses disciplines from history and economics, to politics and geography; within which, sociology is a subset that studies the science of society. The word ‘sociology’ is a hybrid, coined by French philosopher Isidore Auguste Comte in 1830 “from Latin socius ‘associate’ + Greek-derived suffix –logie”; more specifically, “social” as the adjective dates from the 1400s, meaning “devoted to or relating to home life”; and 1560s as “living with others”. The term’s domestic connotation soon accelerated from the realm of the private to the public: “Social Contract” from translations of Rousseau in 1762; “Social Darwinism” and “Social Engineering” introduced by Fisher and Marken in 1877 and 1894; “Social Network” and “Social Media” by the late 20th century from Ted Nelson. Blooming during a high time of the Enlightenment and the rise of the positivist worldview, sociology naturally claims itself to be a science, of scientific methods and empirical investigations. The connotation of –logie has been brilliantly attested by Jonathan Culler:
“Traditionally, Western philosophy has distinguished ‘reality’ from ‘appearance’, things themselves from representations of them, and thought from signs that express it. Signs or representations, in this view, are but a way to get at reality, truth, or ideas, and they should be as transparent as possible; they should not get in the way, should not affect or infect the thought or truth they represent.”
To claim a social study as a science puts forward the question of the relationship between the language that is used to empirically describe and analyse the subject with the subject matter itself. If it should be objectively and rationally portrayed, then the language of mathematics would seem perfect for the job. If we are able to describe the interaction between two or more people using mathematics as a language, then we may begin to write down a partial differential equation and map the variables of it. Algorithms that are inductively trained on evidence-based data do not only seem to capture the present state of such interaction, but seem also able to give critical information in describing the future evolution of the system. This raises the question of computability: what is the limit to social computation? If there is none, then we might as well be a simulation ourselves; so the logic goes that there must be one. To leave an algorithm running without questioning the limits to social computation is like having Borel’s monkey hitting keys at random on a typewriter, or to apply [insert promising algorithm here] arbitrarily for [insert ear-catching grand challenges here].
What’s the hook?
A hook “is a musical idea, often a short riff, passage, or phrase, that is used in popular music to make a song appealing and to catch the ear of the listener”. It is a monumental part of Web 2.0 that takes user attention as a scarce resource and a valuable commodity – an attention economy. Music is an artform that takes time to comprehend; as it plays through time, it accrues value in your attention.
This is one of the most famous hooks of the late 2000s – Empire State of Mind came around the same time as the Web 2.0 boom, just after New York had recovered from the dotcom bubble. The song was like an acoustic montage of the “Eight million stories, out there in the naked’, revealing an underlying urge for social change that was concealed by the boom; just as we see Jay-Z in Times Square on stage under the “big lights that inspired” him rapping: “City is a pity, half of y’all won’t make it”. It was an epoch of R&B, rhythms of cities, of the urban sphere, of the hightech low life. Just the first 15 seconds of Jay-Z’s beat is already enough to teleport a listener to Manhattan, with every bit of romanticism that comes with it. The Rhythms and the Blues constructed a virtual space of narrative and story-telling; such spatial quality taps into the affective experiences of the listener through the ear, revealing the urban condition through its lyrical expression. It is no accident that the 2000s was also a time when the artist / sculptor Isa Genzken began exploring the potential of audio in its visual-sculptural embodiment.
“The ear is uncanny. Uncanny is what it is; double is what it can become; large [or] small is what it can make or let happen (as in laisser-faire, since the ear is the most [tender] and most open organ, the one that, as Freud reminds us, the infant cannot close); large or small as well the manner in which one may offer or lend an ear.” — Jacques Derrida.
An image of a woman’s ear was placed on a facade by Genzken, personifying the building as a listener, hearing what the city has to say. At the same time, “The body is objectified and made into a machine that processes external information”. The ear also symbolises the power of voice that could fill a place with a space: an acoustic space. As much as a place is a location, geographically tagged, and affects our identity and self-association of belonging; a space can be virtual as much as it can be physical. Such a space of social interaction is now being visualised on a facade, and at the same time, it is being fragmented: “To look at a room or a landscape, I must move my eyes around from one part to another. When I hear, however, I gather sound simultaneously from all directions at once: I am at the centre of my auditory world, which envelopes me. … You can immerse yourself in hearing, in sound. There is no way to immerse yourself similarly in sight”. This is perhaps a prelude to augmented virtual reality.
As much as Genzken is interested in the ‘‘exploration of contradictions of urban life and its inherent potential for social change”, Rem Koolhaas shared a similar interest in his belief that it is not possible to live in this age if you don’t have a sense of many contradictory voices.  What the two have in common is their continental European roots and a love for the Big Apple – Genzken titled her 1996 collage book “I Love New York, Crazy City”, and with it paid homage to her beloved city. Delirious New York was written at a time when New York was on the verge of bankruptcy, yet Koolhaas saw it as the Rosetta Stone, and analysed the city as if there had been a plan, with everything starting from a grid. It was Koolhaas’ conviction that the rigor of the grid enabled imagination, despite its authoritative nature: unlike Europe, which has many manifestos with no manifestation, New York was a city with a lot of manifestation without manifesto.
Koolhaas’ book was written with a sense of “critical paranoia” – a surrealist approach that blends together pre-existing conditions and illusions to map the many blocks of Manhattan into a literary montage. The cover of the first edition of the book, designed by Madelon Vriesendorp, perfectly captures the surrealism of the city’s socio-economy at the time: the Art Deco skyscraper Chrysler Building is in bed with the Empire State. Both structures were vying for distinction in the “Race into the Sky” of the 1920s, fueled by American optimism, a building boom, and speculative financing.  Just as the French writer Lautréamont wrote: “Beautiful as the accidental encounter, on a dissecting table, of a sewing machine and an umbrella”, surrealism is a paradigmatic shift of “a new type of surprising imagery replete with disguised sexual symbolism” The architectural surrealism manifested in this delirious city is the chance encounter of capital, disguised as national symbolism – an architectural hook.
Genzken’s sense of scale echoes Koolhaas’ piece on “bigness” in 1995. Her proposal for the Amsterdam City Gate frames and celebrates the empty space, and found manifestation in Koolhaas’ enormous China Central Television’s (CCTV) Beijing headquarters – a building as a city, an edifice of endless air-conditioning and information circularity wrapped in a structured window skin, hugging itself in the air by its downsampled geometry of a mobius loop. Just as Koolhaas pronounced, within a world that tends to the mega, “its subtext is f*** context”. One is strongly reminded of the big data approach to form-finding, perhaps also of the discrete spatial quality coming from Cellular Automata (CA), where the resolution of interconnections and information consensus fades into oblivion, turning data processing into an intelligent, ever mounting aggregation. In the big data–infused era, the scale boundary between architecture and urban design becomes obscured. This highlights our contemporary understanding of complex systems science, where the building is not an individual object, but part of a complex fabric of socioeconomic exchanges.
As Carpo captured in his Second Digital Turn, we are no longer living in Shannon’s age, where compression and bandwidth is of highest value: “As data storage, computational processing power, and retrieval costs diminish, many traditional technologies of data-compression are becoming obsolete … blunt information retrieval is increasingly, albeit often subliminally, replacing causality-driven, teleological historiography, and demoting all modern and traditional tools of story-building and story-telling. This major anthropological upheaval challenges our ancestral dependance on shared master-narratives of our cultures and histories”. Although compression as a skillset is much used in the learning process of the machines for data models, from autoencoders to convolutional neural networks, trends in edge AI and federated learning are displacing value in bandwidth with promises of data privacy – we no longer surrender data to a central cloud, instead, all is kept on our local devices with only learnt models synchronising.
Such displacement of belief in centralised provisions to distributed ownership is reminiscent of the big data-driven objectivist approach to spatial design, which gradually displaces our faith in anything non-discursive, such as norms, cultures, and even religion. John Lagerwey defines religion in its broadest sense as the structuring of values. What values are we circulating in a socio-economy of search engines and pay-per-clicks? Within trends of data distribution, are all modes of centrally-provisioned regulation and incentivisation an invasion of privacy? Genzken’s work in urbanity is like a mirror held up high for us to reflect on our urban beliefs.
Genzken began architecturing a series of “columns” around the same time as her publication of I Love New York, Crazy City. Evocative of skyscrapers and skylines that are out of scale, she named each column after one of her friends, and decorated them with individual designs, sometimes of newspapers, artefacts, and ready-made items that reflect the happenings of the time. Walking amongst them reminds the audience of New York’s avenues and its urban strata, but at 1:500. Decorated with DIY store supplies, these uniform yet individuated structures seem to be documenting a history of the future of mass customization. Mass customisation is the use of “flexible computer-aided manufacturing systems to produce custom output. Such systems combine the low unit costs of mass production processes with the flexibility of individual customization”. As Carpo argued, mass customisation technologies would potentially make economies-of-scale and their marginal costs irrelevant and, subsequently, the division-of-labour unnecessary, as the chain of production would be greatly distributed. The potential is to democratise the privilege of customised design, but how can we ensure that such technologies would benefit social goals, and not fall into the same traps of the attention economy and its consumerism?
Refracted and reflected in Genzken’s “Social Facades” – taped with ready-made nationalistic pallettes allusive of the semi-transparent curtain walls of corporate skyscrapers – one sees nothing but only a distorted image of the mirrored self. As the observer begins to raise their phone to take a picture of Genzken’s work, the self suddenly becomes the anomaly in this warped virtual space of heterotopia.
“Utopia is a place where everything is good; dystopia is a place where everything is bad; heterotopia is where things are different – that is, a collection whose members have few or no intelligible connections with one another.” — Walter Russell Mead 
Genzken’s heterotopia delineates how the “other” is differentiated via the images that have been consumed – a post-Fordist subjectivity that fulfils itself through accelerated information consumption.
The Algorithmic Form
Genzken’s engagement with and interest in architecture can be traced back to the 1970s, when she was in the middle of her dissertation at the academy. She was interested in ellipses and hyperbolics, which she prefers to call “Hyperbolo”. The 70s were a time when a computer was a machine that filled the whole room, and to which a normal person would not have access. Genzken got in touch with a physicist, computer scientist Ralph Krotz, who, in 1976, helped in the calculation of the ellipse with a computer, and plotted the draft of a drawing with a drum plotter that prints on continuous paper. Artists saw the meaning in such algorithmic form differently than scientists. For Krotz, ellipses are conic sections. Colloquially speaking, an egg comes pretty close to an ellipsoid: it is composed of a hemisphere and half an ellipse. If we are to generalise the concept of conic section, hyperbolas also belong to it: if one rotates a hyperbola around an axis, a hyperboloid is formed. Here, the algorithmic form is being rationalised to its computational production, irrelevant of its semantics – that is, until it was physically produced and touched the ground of the cultural institution of a museum.
The 10-meter long ellipse drawing was delivered full size, in one piece, as a template to a carpenter, who then converted it to his own template for craftsmanship. Thus, 50 years ago, Genzken’s work explored the two levels of outsourcing structure symbolic of today’s digital architectural production. The output of such exploration is a visual-sculptural object of an algorithmic form at such an elongated scale and extreme proportion that it undermines not only human agency in its conception, but also the sensorial perception of 2D-3D space. When contemplating Genzken’s Hyperbolo, one is often reminded of the radical play with vanishing points in Hans Holbein’s “The Ambassadors”, where the anamorphic skull can only be viewed at an oblique angle, a metaphor for the way one can begin to appreciate the transience of life only with an acute change of perspective.
When situated in a different context, next to Genzken’s aircraft windows (“Windows”), the Hyperbolo finds association with other streamlined objects, like missiles. Perhaps the question of life and death, paralleling scientific advancement, is a latent meaning and surrealist touch within Genzken’s work, revealing how the invention of the apparatus is, at the same time, the invention of its causal accidents. As the French cultural theorist and urbanist Paul Virilio puts it: the invention of the car is simultaneously the invention of the car crash. We may be able to compute the car as a streamlined object, but we are not even close to being able to compute the car as a socio-cultural technology.
Perhaps the problem is not so much whether the “social” is computable, but rather that we are trying to objectively rationalise something that is intrinsically social. This is not to say that scientific methods to social architecture are in vain; rather the opposite, that science and its language should act as socioeconomic drivers to changes in architectural production. What is architecture? It can be described as what stands at the intersection of art and science – the art of the chief ‘arkhi-’ and the science of craft ‘tekton’ – but the chance encounter of the two gives birth to more than their bare sum. If architecture is neither art nor science but an emergence of its own faculty, it should be able to argue for itself academically as a discipline, with a language crafted as its own, and to debate itself on its own ground – beyond the commercial realm that touches base with ground constraints and reality of physical manifestation, and also in its unique way of researching and speculating, not all “heads in the clouds”, but in fact revealing pre-existing socioeconomic conditions.
It is only through understanding ourselves as a discipline that we can begin to really grasp ways of contributing to a social change, beyond endlessly feeding machines with data and hoping it will either validate or invalidate our ready-made and ear-catching hypothesis. As Carpo beautifully put it:
“Reasoning works just fine in plenty of cases. Computational simulation and optimization (today often enacted via even more sophisticated devices, like cellular automata or agent-based systems) are powerful, effective, and perfectly functional tools. Predicated as they are on the inner workings and logic of today’s computation, which they exploit in full, they allow us to expand the ambit of the physical stuff we make in many new and exciting ways. But while computers do not need theories, we do. We should not try to imitate the iterative methods of the computational toolds we use because we can never hope to replicate their speed. Hence the strategy I advocated in this book: each to its trade; let’s keep for us what we do best.” 
1 A. Bava, “Computational Tendencies – Architecture – e-Flux.” Computational Tendencies, January. 2020. https://www.e-flux.com/architecture/intelligence/310405/computational-tendencies/.
2 R. Bottazzi, Digital Architecture beyond Computers Fragments of a Cultural History of
Computational Design (London: Bloomsbury Visual Arts, 2020).
3 ASSRU, Algorithmic Social Sciences, http://www.assru.org/index.html. (Accessed December 18, 2021)
4 P. Schumacher, Design of Information Rich Environments, 2012.
5 Oxford, “The Home of Language Data” Oxford Languages, https://languages.oup.com/ (Accessed December 18, 2021).
6 Google, “Algorithmic Information Theory – Google Arts & Culture”, Google,
https://artsandculture.google.com/entity/algorithmic-information-theory/m085cq_?hl=en. (Accessed December 18, 2021).
7 Britannica, “Sociology”, Encyclopædia Britannica, inc. https://www.britannica.com/topic/sociology. (Accessed December 18, 2021).
8 Etymonline, “Etymonline – Online Etymology Dictionary”, Etymology dictionary: Definition, meaning and word origins, https://www.etymonline.com/, (Accessed December 18, 2021).
9 J. Culler, Literary Theory: A Very Short Introduction, (Oxford: Oxford University Press, 1997).
10 K. Friston, ”The free-energy principle: a unified brain theory?“ Nature reviews neuroscience, 11 (2),127-138. (2010)
11 J. Covach, “Form in Rock Music: A Primer” (2005), in D. Stein (ed.), Engaging Music: Essays in Music Analysis. (New York: Oxford University Press), 71.
12 Jay-Z. Empire State Of Mind, (2009) Roc Nation, Atlantic
13 J. Derrida, The Ear of the Other: Otobiography, Transference, Translation ; Texts and Discussions with Jacques Derrida. Otobiographies / Jacques Derrida, (Lincoln, Neb.: Univ. of Nebraska Pr., 1985).
15 Kunsthalle Wien, “Kunsthalle Wien #FemaleFool Booklet I’m Isa Genzken the …,” (2014). https://kunsthallewien.at/101/wp-content/uploads/2020/01/booklet_i-m-isa-genzken-the-only-female-fool.pdf?x90478.
16 W. Ong, Orality and Literacy: The Technologizing of the Word, (London: Methuen, 1982)
17 R. Koolhaas, New York délire: Un Manifeste rétroactif Pour Manhattan, (Paris: Chêne, 1978).
18 Kunsthalle Wien, “Kunsthalle Wien #FemaleFool Booklet I’m Isa Genzken the …,” (2014). https://kunsthallewien.at/101/wp-content/uploads/2020/01/booklet_i-m-isa-genzken-the-only-female-fool.pdf?x90478.
19 J. Rasenberger, High Steel: The Daring Men Who Built the World’s Greatest Skyline, 1881 to the Present, (HarperCollins, 2009)
20 Tate, “’L’Enigme D’Isidore Ducasse’, Man Ray, 1920, Remade 1972”, Tate. https://www.tate.org.uk/art/artworks/man-ray-lenigme-disidore-ducasse-t07957, (Accessed December 18, 2021)
21 M. Carpo, ”Big Data and the End of History”. International Journal for Digital Art History, 3: Digital Space and Architecture, 3, 21 (2018)
22 J. Lagerwey, Paradigm Shifts in Early and Modern Chinese Religion a History, (Boston, Leiden: Brill, 2018).
23 Google, “Mass Customization – Google Arts & Culture.” Google, https://artsandculture.google.com/entity/mass-customization/m01k6c4?hl=en (Accessed December 18, 2021).
24 M. Carpo, The Second Digital Turn: Design beyond Intelligence, (Cambridge: MIT, 2017).
25 W.R. Mead, (Winter 1995–1996). “Trains, Planes, and Automobiles: The End of the Postmodern Moment”. World Policy Journal. 12 (4), 13–31
26 U. Loock, “Ellipsoide und Hyperboloide”, in Isa Genzken. Sesam, öffne dich!, exhibition cat. (Whitechapel Gallery, London, and Museum Ludwig, Cologne: Kasper, 2009)
27 S. Baier, “Out of sight”, in Isa Genzken – Works from 1973-1983, Kunstmuseum
28 R. Krotz, H. G. Bock, “Isa Genzken”, in exhibition cat. Documenta 7, Kassel 1982, vol. 1, p. 330-331, vol. 2, p. 128-129
29 A. Farquharson, “What Architecture Isn’t” in Alex Farquharson, Diedrich Diederichsen and Sabine Breitwieser, Isa Genzken (London 2006), 33
30 P. Virilio, Speed and Politics: An Essay on Dromology (New York: Columbia University, 1986).
But let us not have recourse to books for principles which may be found within ourselves. What have we to do with the idle disputes of philosophers concerning virtue and happiness? Let us rather employ that time in being virtuous and happy which others waste in fruitless enquiries after the means: let us rather imitate great examples, than busy ourselves with systems and opinions. … For this reason, my lovely scholar, changing my precepts into examples, I shall give you no other definitions of virtue than the pictures of virtuous men; nor other rules for writing well, than books which are well written.
Jean-Jacques Rousseau, Julie ou la Nouvelle Héloïse, Letter XII (William Kenrick transl., 1784)
Children learn to speak their mother tongues through practice and observation. They don’t need grammar rules. Grammar comes later, when it is taught at school. This shows that we may know a language without knowing its grammar. Grammar is an artificial shortcut to fluency, replacing the lengthy process of learning from life. For a fifteen-year-old high school student struggling to learn German, grammar is indispensable. Yet plenty of native German speakers don’t know declensions by heart and still manage to get their word endings right – in speech as much as in writing.
At a higher level of linguistic practice, literary composition too used to have its own rules – rules that were taught at school. Until the end of the nineteenth century rhetoric was a compulsory subject in most European secondary schools. Rhetoric is the science of discourse. It teaches how to find the arguments of speech, how to arrange them in an orderly manner, and how to dress them with words. Rhetoric teaches how to be clear and persuasive. Seen in this light, rhetoric would seem to be a necessary discipline – indispensable, even. Instead, it no longer features in school and university curricula. France stopped teaching rhetoric in 1885, when French lycées replaced it with the history of classic and modern literature. Nineteenth-century educators seemed to have concluded that, when learning to write, we are better off in the company of literary masterpieces, rather than engaged in the normative study of classical (or modern) rhetoric. A century after Rousseau, Julie-Héloïse’s pedagogical programme quoted above became law.
In times gone by students would have learnt the art of discourse by systematically studying grammar and rhetoric – page after page of rules to be learnt by heart. Today high school students in all European countries are instead obliged to read the masterpieces of their respective national literatures, often ad nauseam. This evidently follows from the assumption that, by reading and re-reading these exemplary works, students will (at some point) learn to write as beautifully as these canonical authors once did. Never mind that nobody knows precisely how and when that almost magic transference, assimilation, and transmutation of talent might occur: grammar has almost completely disappeared from primary school teaching, and rhetoric barely features in higher education – now an intellectual fossil of sorts. Meanwhile, the old art of discourse tacitly lingers on, in business schools, in creative writing and marketing classes. Especially in the latter, the ancient forensic discipline is returned to one of its ancestral functions: that of persuading, even when in the wrong.
For the Humanists of the Quattrocento, the first language to learn was Latin. Not Medieval Latin of course – a corrupt and barbaric but still living language. Renaissance Humanists wanted to speak in the tongue of classical antiquity; they wanted to learn Cicero’s Latin. But Cicero’s Latin is, by definition, a dead language: quite literally so, since it died with Cicero. Cicero also wrote manuals on the art of rhetoric, but the Humanists believed that the best way to learn to write like Cicero was by imitating his way of writing. Well before the Romantics and the Moderns, they found learning from rules unappealing. They preferred to copy the style of Cicero from examples of his work.
The Humanists’ veneration of examples was not limited to languages. Their exemplarism was an épistémè – an intellectual, cultural and social paradigm, deeply inscribed within the spirit of their time. That was their rebellion against the world they grew up in. For centuries the Scholastic tradition had privileged formalism, deductive reasoning, and syllogistic demonstration. The Humanists rejected this “barbarous”, “Gothic” tradition of logic, in favour of their new way of “learning from examples”. The dry and abstract rules of medieval Scholasticism were difficult to handle. Examples, on the other hand, were concrete and tangible. Imitating an example was easier, more pleasurable, and allowed more room for creativity than merely applying rules. This is how, at the dawn of modernity, antiquity was turned from a rule book into an art gallery.
*** *** ***
Like the arts of discourse, the arts of building require schooling. At the height of the Middle Ages, when both Gothic architecture and Scholasticism were at their peak, architectural lore was the preserve of guilds, and its mostly oral transmission was regulated by secretive initiation practices. By contrast, the Humanists pursued a more open strategy – reviving the ancient custom of writing books on building. The first modern treatise, Alberti’s De Re Aedificatoria, deals with the architecture of antiquity, but the structure of Alberti’s discourse was still medieval and Scholastic. Alberti advocates classical architecture as a paragon for all modern building, but Alberti’s antiquity was an abstract model, devoid of any material, visible incarnation. Rather than an atlas of classical buildings, Alberti’s book offers a set of classical design rules – rules for building in the classical way. To put it in more contemporary terms, Alberti formalized classical architecture. Alberti’s rules replace the need to see – let alone imitate – the monuments of classical antiquity. To avoid all misunderstanding, Alberti’s book did not describe any actual ancient monument, either in writing or visually: Alberti’s De Re Aedificatoria originally did not include any illustrations, and Alberti explained that he wanted it that way.
As a commercial venture, Alberti’s De Re Aedificatoria was not a success. Renaissance architects found it easier to skip Alberti’s writings altogether, and go see, touch and learn from the extant magnificence of Roman ruins in person. Moreover, and crucially, as of the early sixteenth century drawings of ancient monuments started to be sold and circulated throughout Europe. Survey drawings in particular, for the first time made available through print, made the laborious ekphrastic and normative mediation of Alberti’s writings all but unnecessary. But models, if beautiful to behold, are not always easy to imitate. Copies will inevitably be more or less successful, depending on the individual talent of each practitioner. By the second or third decade of the sixteenth century imitation itself had become a pedagogic and didactic conundrum.
Not just architectural imitation: writers had the same problem. After all, imitating Cicero is easier said than done. Many rhetoricians in the sixteenth century will strive to transform the practice, skills, and tacit knowledge of literary imitation into a rational, transmissible technique. The modern notion of “method” was born out of sixteenth century rhetoric, but sixteenth century authors were not trying to develop a (scientific) method for making new discoveries; they were trying to develop a (pedagogic) method to better organise and teach what they already knew. Their post-Scholastic, pre-scientific method was essentially a diairetic method – a method of division: all knowledge, they argued, can be partitioned into smaller and smaller units, easier to learn, remember and work with. For sixteenth century scholars, “method” still meant “short cut” – a short cut to knowledge.
Discourse itself can be divided into modular parts: prefaces, arguments, conclusions, formulas and figures, idioms or turns of phrase, sentences, syntagms, words and letters. Sixteenth-century rhetoricians used this divisive technique to invent a new method for literary imitation. On the face of it, Cicero’s style may appear as an ineffable quintessence, but at the end of the day all writing is text, and every text can be broken down into a linear sequence of alphabetical units. Of course, breaking up a text is not a straightforward operation: the parts of speech are held together by syntactic, semantic, and functional relationships. Some of these links can be uncoupled. Others can’t. A text is a heteroclitic, variable cohesion aggregate of parts. Its segments differ in both extension and complexity. Yet even the most sophisticated literary monument can be subdivided into fragments; and once a fragment has been set apart from its compositional context, it can also be reused, reassembled, or recomposed into another text.
In reducing the art of discourse to a citationist technique – by turning ancient texts into a repository of infinitely repeatable citations – sixteenth century rhetoricians invented a new rhetoric. Ancient and modern texts came to be seen as mechanical assemblages of parts. Ancient works could be decomposed into segments, and these segments could then be reassembled to form new works. The smaller the segments, the more fluid or freer the outcome. Ciceronian Latin was an extraordinarily sophisticated and effective instrument of communication, but some modern ideas fundamentally differed from those of Cicero. The citationist method of imitation allowed Renaissance authors to use an old language to express new ideas.
Renaissance architects also needed a rational method for producing modern buildings while imitating classical examples. The greatest structures of antiquity – temples, amphitheatres, thermal baths – were of no use to modernity. Temples, in particular, while representing the pinnacle of classical architecture, had been built to house rituals and represent heathen gods whose worship had long ceased. The entire language of classical architecture had to be adapted for typologies and functions that had no precedents in antiquity. The image of antiquity itself as a building that can be endlessly dismantled and reassembled was a commonplace in the Renaissance. It was also a common practice on many building sites. Architect Sebastiano Serlio would turn this practice into a design theory.
That was no accident. Giulio Camillo, one of the main theorists of the sixteenth century citationist method, had an interest in architecture. He was also a friend of Serlio. The two were supported by the same patrons, and moved in the same circles of Evangelical (and perhaps Nicodemite) inclination. The method of Giulio Camillo’s Neoplatonist rhetoric is well known:
1. Appropriate ancient examples (literary or otherwise) must be selected. The criteria for this selection were a much-disputed matter at the time, and one on which Camillo himself did not dwell.
2. The resulting corpus of integral textual sources must be segmented or divided into parts according to functional or syntactical criteria.
3. This catalogue of dissolved fragments must be sorted, so new users know where to look for the fragments they need.
4. A modern writer (a composer, but also in a sense a compositor: an ideal type-setter) will pick, reassemble and merge, somehow, any number of chosen textual fragments.
Thus new ideas could be expressed through ancient words and phrases – fragments severed from their original context, yet validated by prior use by a recognised “authority”. In Camillo’s view, this compositional technique constituted the inner workings and the secret formula of all processes of imitation. Furthermore, this was a compositional method that could be taught and learnt.
One essential tool in implementing this pedagogical programme was Camillo’s notorious Memory Theatre, a walk-in filing cabinet where all the textual sources (and possibly some of the fragments deriving from them) would have been sorted following Camillo’s own classification system. The whole machine, which included an ingenious information retrieval device, would have been in the shape of an ancient theatre – and it appears that Camillo built at least a wooden model or mock-up of it, in the hope (soon dashed) of selling his precociously cybernetic technology to King Francis I of France.
In a long-lost manuscript (found and published only in 1983) Camillo also explains how the same principles can inform a new method for architectural design. In Camillo’s Neoplatonic hierarchy of ideas, the heavenly logos descends down into reality following seven steps or degrees of ideality. Individuals inhabit the seventh (lowest, sublunar) step; their ascent and crossing of the lunar sky occurs by dint of their separation from the accidents of space and time. In the case of architecture, actual buildings as they exist on earth must be separated from their site to become ideas of the lowest (sixth) grade. This separation of the real from its worldly context results in something similar to what we would today call “building types” – which are buildings in full, except they do not inhabit any given place. These abstract types are then further subdivided into columns and orders (of the five kinds then known: Tuscan, Doric, Ionic, Corinthian, and Composite). The five orders are then broken down into regular geometric volumes, then surfaces, all the way to Euclidian points and lines. On each grade or step, a catalogue of ready-made parts would offer any designer all the components needed to assemble a new building. Thus Camillo’s design method doubles as a shortcut to architectural imitation, and as a universal assembly kit.
A more scholarly trained Neoplatonist philosopher (and a few existed in Camillo’s time) would have objected to some of Camillo’s brutal simplifications, and could have pointed out that his theory had severe epistemic flaws. All the same, Camillo’s architectural method (which its first editor, Lina Bolzoni, dated to around 1530) is almost identical to the plan laid out by Serlio in the introduction to the first instalment of his architectural treatise, published in Venice in 1537. Some of Serlio’s seven grades did not correspond to Camillo’s order: most notably, his atlas of archaeological evidence, the base and foundation of Camillo’s Neoplatonic scaffolding, should have been on the lowest step, but was instead printed as Serlio’s Third Book (likely for commercial reasons). Additionally, one of the seven books in Serlio’s original plan, his revolutionary Sixth Book, on Dwellings for all Grades of Men, was written but never published – at least, not until 1966. Serlio also wrote an additional, Extraordinary Book (literally, a book out of the original order) – a cruel, sombre joke disguised as a book, which Serlio bequeathed to posterity shortly before dying, poor and dejected in his self-imposed French exile.
Regardless of some factual discrepancies, Serlio’s compositional method is ostensibly the same as Camillo’s. Architecture’s exemplary models are selected, and then fragmented. These fragments are sorted and classified at different levels or grades of dissolution. Instructions for their reassembly are then provided, together with examples of successful new compositions. The pivot of the whole system was the book on the five architectural orders, which Serlio published first (albeit titled Fourth Book to comply with the general plan): a catalogue of stand-alone constructive parts (columns, capitals, bases, entablatures and mouldings), destined for identical reproduction in print, in scaled drawings, and in buildings of any type. In Serlio’s method, this was the main offspring of architectural “dissolution” (or disassembling), and the basic ingredient of architectural design, i.e. re-composition. Pagan idols had to be broken down; only their fragments could be used, purified ingredients in the building of a new Christian architecture.
All the way, Serlio was aware of, and attuned to, the purpose and limits of his architectural method. Serlio turned architectural design into an assemblage of ready-made modular components. These were not actual spolia, but compositional design units, part to a universal combinatory grammar and destined for identical replication. Giulio Camillo’s rhetoric reduced the imitation of Cicero’s style, hence all literary composition, to a cut-and-paste method of collage and citation. Serlio’s treatise did the same for architecture. His theory of the orders was the keystone of the entire process. Serlio couldn’t standardise the building site (that would have made no sense in the sixteenth century), but he could standardise architectural drawings and design.
Serlio knew full well that his simplified, almost mechanical approach to design would entail a decline in the general quality of architecture. Many critics across the centuries have indeed frowned at the models and projects shown in his Seven Books. Serlio’s designs have often been seen as repetitive, banal, ungainly or chunky; lacking in inspiration and genius. But Serlio did not write for geniuses. His treatise was a pedagogical work, not an architectural one. As Serlio tirelessly reminds the reader, his method is tailored to “every mediocre”: to the “mediocre architect” – the average, middling designer. Today we might say that Serlio’s treatise aimed at creating an intermediate class of building professionals. Michelangelo and Raphael had no need for “a brief and easy method” that turned architectural invention into cut-and-paste, collage and citation.
Knowledge can be taught, not genius. Serlio’s pedagogical structure and design method were parts of an overarching ideological project. Serlio’s method promises uniform and predictable architectural standards. These are perhaps banal, or monotonous, but that’s the price one pays to make “architecture easy for everyone”. And it is a price Serlio was willing to pay. Serlio’s concern was the average quality of building, not the artistic value of a few outstanding monuments. This was a most unusual choice for an artist of the Italian Renaissance – an iconoclastic, almost revolutionary stance. Serlio’s worldview was not one in which the misery of the many was contrasted by the magnificence of a few. Serlio pursued the uniform, slightly boring repetitiveness of a productive, “mediocre” multitude. This was an ideological project, but also a social project, ripened in the cultural context of the early protestant Reformation. It is a position that evokes and preludes well-known categories of modernity.
* Footnote to this translation
This is a translation of the introduction to my book Metodo e Ordini nella Teoria Architettonica dei Primi Moderni (Geneva: Droz, Travaux d’Humanisme et Renaissance, 1993), edited, abridged, and adapted for clarity, but not updated. That book in turn derived from my PhD dissertation, supervised by Joseph Rykwert, researched and written between 1984 and 1989, and defended in the spring of 1990. Heavily influenced by Françoise Choay’s La Règle et le Modèle and by works of literary criticism by Terence Cave (The Cornucopian Text), Antoine Compagnon (La seconde main ou le travail de la citation), and Marc Fumaroli (L’âge de l’éloquence), all published between 1979 and 1980, my enquiry on the use of visual citations in Renaissance architectural design was evidently in the spirit of the time: post-modern architects in the 80s were passionate about citations (or the recycling of precedent, otherwise known as reference, allusion, collage and cut-and-paste); they were equally devoted to architectural history, and particularly to the history of Renaissance classicism. My aim then was to bridge the gap between those two sources of PoMo inspiration, showing that Renaissance architecture was itself, quintessentially, citationist. How could it have been otherwise, since the main purpose of Renaissance architects was to revive, literally, the buildings of classical antiquity – piece by piece? Thanks to the first studies of Lina Bolzoni on the sulphurous Renaissance philosopher and magician Giulio Camillo, and to my then girlfriend, who was studying Renaissance Neoplatonism (and is today a known specialist of that arcane science), I soon found evidence of an extraordinary link – biographical, ideological, and theoretical – between Giulio Camillo and Sebastiano Serlio, and I wrote a PhD dissertation to explain the transference of the citationist method from Bembo’s Prose to Camillo’s Theatre to Serlio’s Seven Books – and ultimately to Serlio’s architecture.
Unfortunately, in the process, I also found out that the citationist method in the 16th century was a tool and vector of modernity. It was a mechanical method, made to measure for the new technology of printing; it was also in many ways a harbinger of the scientific revolution that would soon follow. Besides, the citationist method was more frequently adopted by Evangelical and Protestant thinkers (particularly Calvinist), and it was condemned by the Counter-Reformation. None of this would have pleased the PoMo architects and theoreticians who were then my main interlocutors.
Fortunately for me, they never found out. When my book was published, in 1993, the tide of PoMo citationism was already receding. Investigating the sources of citationism was no longer an urgent matter for architects and designers. My book was published in Italian, in an austere collection of Renaissance studies – few architects would have known about it, let alone read it. It received some brutally disparaging reviews, as due, by some of Tafuri’s acolytes, because they thought, without reading my book, or misreading it, that I was bringing water to the PoMo mill. I wasn’t. But at that point that was irrelevant. We had all already moved on.
I was pleasantly surprised when, a few years ago, Jack Self commissioned this translation for publication in Real Review (the translation, by Fabrizio Ballabio, was soon thereafter partially republished in Scroope, the journal of the Cambridge School of Architecture, at the request of Yasmina Chami and Savia Palate); and I was of course more than happy when my colleague Alessandro Bava asked me to review it for publication in the B-Pro journal of Bartlett School of Architecture. As we all know, collage and citation are becoming trendy again in some architectural circles – for reasons quite different from those of the late structuralists and early PoMos that were my mentors when I was a student. I have somewhat mixed feelings about the current, post-digital revival of collaging, but I would be happy to restart a discussion we briefly adjourned a generation ago.
Mario Carpo (March 2022)
Metodo e Ordini nella Teoria Architettonica dei Primi Moderni. Alberti, Raffaello, Serlio e Camillo (Geneva: Droz, 1993). 226 pages. Travaux d’Humanisme et Renaissance, 271
“Citations, Method, and the Archaeology of Collage”. Real Review, 7 (2018): 22-30, transl. by Fabrizio Ballabio and by the author; partly republished in Scroope, Cambridge Architectural Journal, 28 (2019): 112-119
Part relationships play an important role in architecture. Whether an aspect of a Classical order, a harmonious joining of building components, a representation of space, a partition of spaces, or as a body that separates us and identifies us as individuals. From the very outset, every form of architecture begins with an idea of how parts come together to become a whole and an understanding of how this whole relates to other parts. Architecture first composes a space as a part of a partitioning process well before defining a purpose, and before using any geometry.
The sheer performance of today’s computational power makes it possible to form a world without a whole, without any third party or a third object. Ubiquitous computing fosters peer-to-peer or better part-to-part exchange. It is not surprising then that today’s sharing represents an unfamiliar kind of partiality. From distributive manufacturing to the Internet of Things, new concepts of sharing promise systematic shifts, from mass-customisation to mass-individualisation: the computational enabled participations are foundational. It is no longer the performance or mode of an algorithm that drives change but its participatory capacities. From counting links, to likes, to seats, to rooms: tools for sharing have become omnipresent in our everyday lives. Thus, that which is common is no longer negotiated but computed. New codes – not laws or ideologies – are transforming our cities at a rapid pace, but what kind of parthood is being described? How does one describe something only through its parts today? To what extent do the automated processes of sharing differ from the partitioning of physical space? How can we add, intervene and design such parts through architecture?
The relationship between parts and their whole is called Mereology. In this issue of Prospectives, mereology’s theories and the specifics of part-relations are explored. The differences between parts and the whole, the sharing of machines and their aesthetics, the differences between distributive and collective, their ethical commitments, and the possibilities of building mereologies are discussed in the included articles and interviews.
Just as mereology describes objects from their parts, this issue is partial. It is not a holistic proposal, but a collection of positions. Between philosophy, computation, ecology and architecture, the texts are reminders that mereologies have always been part of architecture. Mereology is broadly a domain that deals with compositional possibilities, relationships between parts. Such an umbrella – analogue to morphology, typology, or topology – is still missing in architecture. Design strategies that depart part-to-part or peer-to-peer are uncommon in architecture, also because there is (almost) no literature that explores these topics for architectural design. This issue hopes to make the extra-disciplinary knowledge of mereology accessible to architects and designers, but also wishes to identify links between distributive approaches in computation, cultural thought and built space.
The contributions gathered here were informed by research and discussions in the Bartlett Prospectives (B-Pro) at The Bartlett School of Architecture, UCL London from 2016 to 2019, culminating in an Open Seminar on mereologies which took place on 24 April 2019 as part of the Prospectives Lecture Series. The contributions are intended as a vehicle to inject foundational topics such as mereology into architectural design discourse.
This compilation starts with Giorgio Lando’s text “Mereology and Structure”. Lando introduces what mereology is for philosophers, and why philosophers discuss mereological theses, as well as disagree one with another about them. His text focuses in particular on the role of structure in mereology outlining that from a formal point of view part relations are freed from structure. He argues that independence from structure might be the identifying link between mereology and architecture. The second article “From Partitioning to Partaking” is a plea for re-thinking the city. Daniel Koehler’s essay points to the differences between virtual and real parts. Koehler observes a new spatial practice of virtual representations that render previous models of urban governance obsolete. He argues that the hyper-dimensional spaces of a big data-driven economy demand a shift from a partitioning practice of governance to more distributed forms of urban design. In “Matter versus Parts: The Immaterialist Basis of Architectural Part-Thinking” Jordi Vivaldi Piera highlights the revival of matter in parallel to the discrete turn in contemporary discourses on experimental architecture. The essay gravitates around the notion of part-thinking in association with the notion of form. Fluctuating between continuous and discrete, the text sets out requirements for radical part-thinking in architecture. As a computational sociologist, David Rozas illustrates the potential of decentralised technologies for democratic processes at the scale of neighborhood communities. After an introduction to models of distributed computation, “Affordances of Decentralised Technologies for Commons-based Governance of Shared Technical Infrastructure” draws analogies to Elinor Ostrom’s principles of commons governance and how those can be computationally translated, turning community governance into fully decentralised autonomous organisations.
Departing from the Corbusian notion of a ‘machine for living’, Sheghaf Abo Saleh defines a machine for thinking. In “When Architecture Thinks! Architectural Compositions as a Mode of Thinking in the Digital Age” Abo Saleh states that the tectonics of a machine that thinks is brutal and rough. As a computational dialogue, she shows how roughness can enable posthumanism which, in her case, turns “tempered” parts into a well-tempered environment. Ziming He’s entry point for “The Ultimate Parts” is the notion of form as the relations between parts and wholes. He’s essay sorts architectural history through a mereological analysis, proposing a new model of part-to-part without wholes. Shivang Bansal’s “Towards a Sympoietic Architecture: Codividual Sympoiesis as an Architectural Model” investigates the potential of sympoiesis. By extending Donna Haraway‘s argument of “tentacular thinking” into architecture, the text shifts focus from object-oriented thinking to parts. Bansal argues for the limits of autopoiesis as a system and conceptualises spatial expressions of sympoiesis as a necessity for an adaptive and networked existence through “continued complex interactions” among parts.
Merging aspects of ‘collective’ and ‘individuality,’ in “Codividual Architecture within Decentralised Autonomous System” Hao Chen Huang proposes a new spatial characteristic that she coins as the “codividual”. Through an architectural analysis of individual and shared building precedents, Huang identifies aspects of buildings that merge shared and private features into physical form. Anthony Alviraz’s paper “Computation Within Codividual Architecture” investigates the history and outlook of computational models into architecture. From discrete to distributed computation, Alviraz speculates on the implications of physical computation where physics interactions overcome the limits of automata thinking. In “Synthesizing Hyperumwelten”, Anna Galika transposes the eco-philosophical concept of an HyperObject into a “Hyperumwelt”. While the Hyperobject is a closed whole that cannot be altered, a Hyperumwelt is an open whole that uses objects as its parts. The multiple of a Hyperumwelt offers a shift from one object’s design towards the impact of multiple objects within an environment.
Challenging the notion of discreteness and parts, Peter Eisenman asks in the interview “Big Data and the End of Architecture Being Distant from Power” for a definition of the cultural role of the mereological project. Pointing to close readings of postmodern architecture that were accelerated by the digital project, Eisenman highlights that the demand for a close reading is distanced from the mainstream of power. The discussion asks: ultimately, what can an architecture of mereology critique? The works of Herman Hertzberger are an immense resource on part-thinking. In the interview “Friendly Architecture: In the Footsteps of Structuralism”, Herman Hertzberger explains his principle of accommodation. When building parts turn into accommodating devices, buildings turn into open systems for staging ambiguity.**
The issue concludes with a transcript from the round table discussion at the Mereologies Open Seminar at The Bartlett School of Architecture on 24 April 2019.
The contributions evolved within the framework of Bartlett Prospectives (B-Pro) at The Bartlett School of Architecture, UCL. I want to thank Frédéric Migayrou for his vision, commitment and long years of building up a research program, not only by architecture but through computation. I would like to thank Roberto Bottazzi for the years of co-organising the Prospectives Lecture Series, where plenty of the discussions that form the backbone of this issue took place. Thanks to Mario Carpo for raising the right question at the right time for so many people within the program, thanks to Andrew Porter for enabling so many events, to Gilles Retsin, for without the discrete there are no parts, Mollie Claypool for the editing and development of Prospectives journal, and Vera Buehlmann, Luciana Parisi, Alisa Andrasek, Keller Easterling, Matthew Fuller, John Frazer, Philippe Morel, Ludger Hovestadt, Emmanuelle Chiappone-Piriou, Jose Sanchez, Casey Rehm, Tyson Hosmer, and Jordi Vivaldi Piera for discussions and insights.
I want to thank Rasa Navasaityte, my partner in Research Cluster 17 at B-Pro, for driving the design research. Thank you for the research contributed by the researchers and tutors: Christoph Zimmel, Ziming He, Anqi Su, Sheghaf Abo Saleh, and to all participants, specifically to highlight: Genmao Li, Zixuan Wang, Chen Chen, Qiming Li, Anna Galika, Silu Meng, Ruohan Xu, Junyi Bai, Qiuru Pu, Anthony Alviraz, Shivang Bansal, Hao-Chen Huang, Dongxin Mei, Peiwen Zhan, Mengshi Fu, Ren Wang, Leyla El Sayed Hussein, Zhaoyue Zhang, Yao Chen, and Guangyan Zhu.
The issue includes articles that evolved from thesis reports conducted in the following clusters: Ziming He from Research Cluster 3 tutored by Tyson Hosmer, David Reeves, Octavian Gheorghiu, and Jordi Vivaldi in architecture theory. Sheghaf Abo Saleh, Anthony Alvidrez, Shivang Bansal, Anna Galika, Hao Chen Huang from Research Cluster 17 tutored by Daniel Koehler and Rasa Navasaityte. If not indicated directly, the featured images, graphics of this issue are by Daniel Koehler, 2020.
In this article I will illustrate affordances of decentralised technologies in the context of commons governance. My aim is to summarise the conversation around the lecture “When Ostrom Meets Blockchain: Exploring the Potentials of Blockchain for Commons Governance” I gave in the Mereologies Open Seminar organised by The Bartlett School of Architecture at University College London on 25th April 2019. I will also extend the conversation by providing a concrete example of such affordances in the context of a community network.
What is Blockchain? Three Key Concepts around Decentralised Technologies
In 2008, an anonymous paper presented Bitcoin: the first cryptocurrency based purely on a peer-to-peer system. For the first time, no third parties were necessary to solve problems such as double-spending, thanks to cryptography. The solution was achieved through the introduction of a data structure known as a blockchain. In simple terms, a blockchain can be understood as a distributed ledger. Distributed refers to a technical property of a system in which certain components are located on different computers connected through a network. The blockchain, in this sense, can be thought of as a “decentralised book” in which agreed transactions can be stored in a set of distributed computers. Data, such as the history of monetary exchanges generated by using cryptocurrencies, can be stored in a blockchain. The key aspect resides in the fact that there is no need to trust a third party, such as a bank server, to store that information.
Nakamoto’s article opened what is considered to be the first generation of blockchain technologies. This generation, up to approximately 2013, includes Bitcoin and a number of crypto-currencies that appeared after it. The second generation, approximately from 2014 onwards, is the extension of these blockchains with capabilities beyond currencies in the form of automatic agreements or smart contracts. Smart contracts can be understood as distributed applications which encode clauses that are automatically enforced and executed without the need for a central authority. They can be employed, for example, to enable the execution of code to provide certifications, such as obtaining a diploma or a registry of lands, according to previously mutually agreed rules. Again, the novel aspect here is the fact that the execution of such rules, in the form of computer instructions, is distributed across a large number of computers without the need of a central point of control.
Complex sets of smart contracts can be developed to make it possible for multiple parties to interact with each other. This fostered the emergence of the last of the concepts I will introduce around decentralised technologies: Decentralised Autonomous Organisations (DAO). A DAO is a self-governed organisation in which interactions between the members of the organisation are mediated by the rules embedded in the DAO code. These rules are sets of smart contracts that encode such interactions. The rules embedded in the code are automatically enforced by the underlying technology, the blockchain, in a decentralised manner. DAOs could, for example, hire people to carry out certain tasks or compensate them for undertaking certain action. Overall, this can be understood as analogous to a legal organisation, with legal documents – bylaws – which define the rules of interaction among members. The development of DAOs has been, unsurprisingly, significantly popular around financial services. However, DAOs could be used to provide a wide variety of services or management of resources in a more diverse range of areas. A more artistic example of a DAO is the Plantoid project, a sculpture of a plant, which can hire artists to physically modify the sculpture itself according to the rules collectively agreed in the smart contracts encoded in it.
All of these potentials of decentralised technologies represent an emerging research field. Together with other colleagues of the EU project P2PModels, we are exploring some of these potentials and limitations in the context of the collaborative economy and, more precisely, on some of the models emerging around Commons-Based Peer Production.
Collaborative Economy and Commons-Based Peer Production
The collaborative economy is a growing socio-economic phenomenon in which individuals produce, exchange and consume services and goods, coordinating through online software platforms. It is an umbrella concept that encompasses different initiatives and significantly different forms are emerging; there are models where large corporations control the platform, thus ensuring its technologies and the knowledge held therein are proprietary and closed. Uber, a riding service, and AirBnB, a short-term lettings service, are perhaps the most well-known examples of such initiatives. They differ from models that revolve around Commons-Based Peer Production (CBPP), where individuals produce public goods by dispensing with hierarchical corporate structures and cooperating with their peers. In these models, participants of the community govern the assets, freely sharing and developing technologies. Some of the most well-known examples of the initiatives around such commons-based models are Wikipedia and GNU/Linux, a Free/Libre Open Source Software (FLOSS) operating system. Commons-based models of the collaborative economy are, however, extending to areas as broad as open science, urban commons, community networks, peer funding and open design.
Three main characteristics are salient in the literature on CBPP. Firstly, CBPP is marked by decentralisation, since authority resides in individual agents rather than a central organiser. Secondly, it is commons-based since CBPP communities make frequent use of common resources. These resources can be material, such as in the case of 3D printers shared in small-scale workshops known as Fab Labs; or immaterial, such as the wiki pages of Wikipedia or the source code in a FLOSS project. Thirdly, non-monetary motivations are prevalent in the community. These motivations are, however, commonly intertwined with extrinsic motivations resulting in a wide spectrum of forms of value operating in CBPP communities, beyond monetary value.
Guifi.net: An Example of a CBPP Community in Action
In order to extend the discussion of the affordances of decentralised technologies in CBPP, I will employ Guifi.net as an illustrative example. Guifi.net is a community network: a participatory project whose goal is to create a free, open and neutral telecommunications network to provide access to the Internet. If you are reading this article online, you might be accessing it through a commercial Internet Service Provider. These are the companies which control the technical infrastructure you are using to connect to the Internet. They manage this infrastructure as a private good. The Guifi.net project, instead, manages this infrastructure as a commons. In other words, Guifi.net is organised around a CBPP model, in which the network infrastructure is governed as a common good. Over the past 16 years, participants of Guifi.net have developed communitarian rules, legal licenses, technological tools and protocols which are constantly negotiated and implemented by the participants.
I have chosen to discuss the potentialities of blockchain drawing on Guifi.net, a community network, for two main reasons. Firstly, the most relevant type of commons governed in this case is shared infrastructure, such as fibre optic and routers. The governance of rival material goods, in contrast to the commons governance of non-rival goods such as source code or wiki pages, better matches the scope of the conversations which emerged during the symposium around architecture of the commons and the role played by participatory platforms. Secondly, Guifi.net provides a large and complex case of governance of shared infrastructure. The growth experienced by Guifi.net’s infrastructure and community since the first pair of nodes were connected in a rural region of Catalonia in 2004 is significant. In their study of the evolution of governance in Guifi.net, Baig et al. reported a network infrastructure consisting of more than 28,500 operational nodes which cover a total length of around 50,000 km of links that are connected to the global Internet. This study refers to the period 2005–2015. The latest statistics reported by Guifi.net state that there are more than 35,000 operational nodes and 63,000 km of links. Beyond the infrastructure, the degree of participation in the community is also significant: more than 13,000 registered participants up to 2015, according to the aforementioned study, and more than 50,000 users of this community network connect on a day to day basis, as reported by the community at present. Thus, Guifi.net provides a suitable scenario for the analysis of the affordances of decentralised technologies for commons governance.
Ostrom’s Principles and Affordances of Decentralised Technologies for Commons Governance
How do communities of peers manage to successfully govern common resources? The study of the organisational aspects of how common goods might be governed was traditionally focussed on the study of natural resources. This commons dilemma was explored by Hardin in his influential article “The Tragedy of the Commons”, whose ideas became the dominant view. In this article, Hardin states how resources shared by individuals acting as homo economicus (out of self-interest in order to maximise their own benefit) results in the depletion of the commons. The individuals’ interests enter into conflict with the group’s, and because they act independently according to their short-term interests, the result of the collective action depletes the commons. As a consequence, in order to avoid this logic – “If I do not use it, someone else will”, which is not sustainable – it was necessary to manage these commons through either private ownership or centralised public administration.
Later on, Nobel laureate researcher Elinor Ostrom questioned and revisited “The Tragedy of the Commons”. In her work, she showed how under certain conditions commons can indeed be managed in a sustainable way by local communities of peers. Her approach took into account that individual agents do not operate in isolation, nor are they driven solely by self interest. Instead, she argued that communities communicate to build processes and rules, with different degrees of explicitation, that ensure their sustainability. This hypothesis was supported by a meta-analysis of a wide range of case studies, and has been confirmed in subsequent research. As part of this work, she identified a set of principles for the successful management of these commons, which has also been subsequently applied to the study of collaborative communities whose work is mediated by digital platforms, such as Wikipedia and FLOSS communities:
1. Clearly defined community boundaries: in order to define who has rights and privileges within the community.
2. Congruence between rules and local conditions: the rules that govern behaviour or commons use in a community should be flexible and based on local conditions that may change over time. These rules should be intimately associated with the commons, rather than relying on a “one-size-fits-all” regulation.
3. Collective choice arrangements: in order to best accomplish congruence (with principle number 2), people who are affected by these rules should be able to participate in their modification, and the costs of alteration should be kept low.
4. Monitoring: some individuals within the community act as monitors of behaviour in accordance with the rules derived from collective choice arrangements, and they should be accountable to the rest of the community.
5. Graduated sanctions: community members actively monitor and sanction one another when behaviour is found to conflict with community rules. Sanctions against members who violate the rules are aligned with the perceived severity of the infraction.
6. Conflict resolution mechanisms: members of the community should have access to low-cost spaces to resolve conflicts.
7. Local enforcement of local rules: local jurisdiction to create and enforce rules should be recognised by higher authorities.
8. Multiple layers of nested enterprises: by forming multiple nested layers of organisation, communities can address issues that affect resource management differently at both broader and local levels.
What kind of affordances do decentralised technologies offer in the context of commons governance and, more concretely, with regards to Ostrom’s principles? Together with other colleagues, we have identified six potential affordances to be further explored.
Firstly, tokenisation. This refers to the process of transforming the rights to perform an action on an asset into a transferable data element (named token) on the blockchain. For example, tokens can be employed to provide authorisation to access a certain shared resource. Tokens may also be used to represent equity, decision-making power, property ownership or labour certificates.
Secondly, self-enforcement and formalisation of rules. These affordances refer to the process of embedding organisational rules in the form of smart contracts. As a result, there is an affordance for the self-enforcement of communitarian rules, such as those which regulate monitoring and graduated sanctions, as reflected in Ostrom’s principles 4 and 5. This encoding of rules also implies a formalisation, since blockchain technologies require these rules to be defined in ways that are unambiguously understood by machines. In other words, the inherent process of explicitation of rules related to the use of distributed technologies also provides opportunities to make these rules more available and visible for discussion, as noted in Ostrom’s principle 2.
Thirdly, autonomous automatisation: the process of defining complex sets of smart contracts which may be set up in such a way as to make it possible for multiple parties to interact with each other without human interaction. This is analogous to software communicating with other software today, but in a decentralised manner. DAOs are an example of autonomous automatisation as they could be self-sufficient to a certain extent. For instance, they could charge users for their services.
Fourthly, decentralised technologies offer an affordance for the decentralisation of power over the infrastructure. In other words, they can facilitate processes of communalising the ownership and control of the technological artefacts employed by the community. They do this through the decentralisation of the infrastructure they rely on, such as collaboration platforms employed for coordination.
Fifthly, transparency: for the opening of organisational processes and the associated data, by relying on the persistency and immutability properties of blockchain technologies.
Finally, decentralised technologies can facilitate processes of codification of a certain degree of trust into systems which facilitate agreements between agents without requiring a third party. Figure 1 below provides a summary of the relationships between Elinor Ostrom’s principles and the aforementioned affordances.
These congruences allow us to describe the impact that blockchain technologies could have on governance processes in these communities. These decentralised technologies could facilitate coordination, help to scale up commons governance or even be useful to share agreements and different forms of value amongst various communities in interoperable ways, as shown by Pazaitis et al.. An example of how such affordances might be explored in the context of CBPP can be found in community networks such as Guifi.net.
A DAO for Commons Governance of Shared Technical Infrastructure
Would it be possible to build a DAO that might help to coordinate collaboration and scale up cooperative practices, in line with Ostrom’s principles, in a community network such as Guifi.net? First of all, we need to identify the relationship between Ostrom’s principles and Guifi.net. We can find, indeed, a wide exploration of the relationship between Ostrom’s principles and the evolution in the self-organisational processes of Guifi.net in the work of Baig et al.. They document in detail how Guifi.net governs the infrastructure as a commons drawing on these principles, and provide a detailed analysis of the different components of the commons governance of the shared infrastructure in Guifi.net. Secondly, we need to define an initial point of analysis, and tentative interventions, in the form of one of the components of this form of commons governance. From all of these components, I will place the focus of analysis on the economic compensation system. The reason for selecting this system is twofold. On the one hand, it reflects the complexity behind commons governance and, thus, allows us to illustrate the aforementioned principles in greater depth. Secondly, it is an illustrative example of the potential of blockchain, as we shall see, to automatise and scale up various cooperative processes.
The economic compensation system of Guifi.net was designed as a mechanism to compensate imbalances in the uses of the shared infrastructure. Professional operators, for example, are requested to declare the expenditures and investments in the network. In alignment with Ostrom’s principle number 4, the use, expenditure and investments of operators are monitored, in this case by the most formal institution which has emerged in Guifi.net: the Guifi.net Foundation. The Foundation is a legal organisation with the goal to protect the shared infrastructure and monitor compliance with the rules agreed by the members of the community. The community boundaries, as in Ostrom’s principle number 1, are clearly defined and include several stakeholders. Different degrees of commitment with the commons were defined as collective choice arrangements (principle number 3). These rules are, however, open to discussion through periodic meetings organised regionally, and adapted to the local conditions, in congruence with principle number 2. If any participant, such as an operator, misuses the resources or does not fulfill the principles, the individual is subject to graduated sanctions, in alignment with principle number 5. As part of the compensation system, compensation meetups are organised locally to cope with conflict resolution, in congruence with principle 6. Principles 6 and 7 are also clearly reflected in the evolution of the governance of Guifi.net, although they are more closely associated with scalability.
The compensation DAO could be formed by a set of local DAOs, whose rules are defined and modifiable exclusively by participants holding a token which demonstrates they belong to this node. These local DAOs could be deployed from templates, and could be modified at any point as a result of a discussion at the aforementioned periodic meetings held by local nodes and in congruence with the local conditions. Among the rules of the smart contracts composing these DAOs, participants may decide to define the different factors that are considered when discussing the local compensation system arrangements, as well as graduated sanctions in case of misuse of the common goods. These rules might be copied and adapted by some of the nodes facilitating the extension of the collaborative practices.
Some of the settings of these local DAOs could be dependent on a federal compensation DAO that defines general aspects. A mapping of the current logic could consist of reaching a certain degree of consensus between the participants in all of the nodes, but having this process approved by the members of the Foundation, who would hold a specific token. Examples of general aspects regulated by the federal DAO are the levels of commitment towards the commons of each operator, which is currently evaluated and monitored manually by the Foundation. General aspects such as this could be automatised in several ways therefore moving from manual assignations by the Foundation, as is currently the case, to automatically assigned tokens depending on the communitarian activities tracked in the platform. This is an example of a possible intervention to automatise certain collaborative practices assuming the current structure. Figure 1 below provides an overview of a preliminary design of a DAO for a compensation system mapping the current logics.
More disruptive tentative interventions could consist of the implementation of more horizontal governance logics which allow modifications of the rules at a federal level or to transform the rules that regulate the monitoring. These interventions, however, should be carefully co-designed together with those who participate in the day-to-day of these collectives. Our approach states that the development of decentralised tools which support commons governance should be undertaken as a gradual process to construct situated technology, with an awareness of the cultural context and aiming to incorporate particular social practices into the design of these decentralised tools.
This basic example of a DAO illustrates, on the one hand, the relationship with Ostrom’s principles: monitoring mechanisms, local collective choice arrangements, graduated sanctions and clear boundaries. These principles are sustained by the aforementioned affordances of blockchain for commons governance. For example, tokenisation with regards to providing permission as to who has the ability to participate in the choices locally and at a federal level and how, as well as the certification of the level of commitment to the commons; monitoring of the expenditures and reimbursements through the transparency provided by the blockchain; self-enforcement, formalisation and automatisation of the communitarian rules in the form of smart contracts. Another, more general, example of this is the increment in the degree of decentralisation of power over the platform because of the inherent decentralised properties of the technology itself. In this way, this could result in a partial shift of power over the platform from the Foundation towards the different nodes formed by the participants. Furthermore, as discussed, the fact that such rules are encoded in the form of configurations of smart contracts could facilitate the extension of practices and the development of new nodes; or even the deployment of alternative networks capable of operating as the former network, and reusing and adapting the encoded rules of the community while still using the shared infrastructure. Overall, further research of the role of decentralised technologies in commons governance offers, in this respect, a promising field of experimentation and exploration of the potential scalability of cooperative dynamics.
Discussion and Concluding Remarks
In this article I provided an overview and discussed an example of the affordances of blockchain technologies for commons governance. Concretely, I described such potentialities drawing on the example of a DAO to automatise some of the collaborative processes surrounding the compensation system of a community network: Guifi.net. Throughout this example, I aimed to illustrate, in more detail, the affordances of blockchain for commons governance which I presented during the symposium. The aim of this example is to illustrate how blockchain may facilitate the extension and scaling up of the cooperation practices of commons governance. Further explorations, more closely related to the architecture field, could explore the discussed affordances for commons governance with discrete design approaches that provide participatory frameworks for collective production. In this respect, decentralised technologies offer opportunities of exploration to tackle challenges such as those identified by Sánchez to define ways to allocate ownership, authorship and distribution of value without falling into extractivist practices.
A better understanding of the capabilities of blockchain technologies for commons governance will require, however, further empirical research. Examples of research questions which need to be addressed are those with regards to the boundaries of the discussed affordances. For example, with regards to tokenisation and formalisation of rules: which aspects should remain in/off the blockchain, or furthermore completely in/out of code?
Overall, CBPP communities provide radically differing values and practices when compared with those in markets. In this respect, the study of the potentialities and limitations of blockchain technologies in the context of the governance of CBPP communities offers an inspiring opportunity to take further steps on a research journey that has only just begun.
 S. Nakamoto,“Bitcoin: A Peer-to-Peer Electronic Cash System” (2008).
 M. Swan, Blockchain: Blueprint for a New Economy (Sebastopol, CA, USA: O’Reilly, 2015).
 N. Szabo, ”Formalizing and Securing Relationships on Public Networks, First Monday, 2, 9 (1997).
 See, for example, https://digix.global: a cryptocurrency backed by bars of gold in which the governance is mediated by a DAO, last accessed on 24th July 2019.
 See http://www.okhaos.com/plantoids/, last accessed on 24th July 2019.
 See https://p2pmodels.eu, last accessed on 2nd July 2019.
 Y. Benkler, The Wealth of Networks: How Social Production Transforms Markets and Freedom (2006); M. Bauwens, “The Political Economy of Peer Production,” CTheory 1, 12 (2005).
 M. Fuster-Morell, J. L. Salcedo, and M. Berlinguer. “Debate About the Concept of Value in Commons-Based Peer Production,” Internet Science (2016); Bauwens, Michel, and Alekos Pantazis. 2018. “The Ecosystem of Commons-Based Peer Production and Its Transformative Dynamics.” The Sociological Review, 66, 2 (2016), 302–19.
 V. Kostakis and M. Papachristou, “Commons-Based Peer Production and Digital Fabrication: The Case of a RepRap-Based, Lego-Built 3D Printing-Milling Machine” (2013); V. Niaros, V. Kostakis, and W. Drechsler, “Making (in) the Smart City: The Emergence of Makerspaces,” Telematics and Informatics (2017).
 A. Arvidsson, A. Caliandro, A. Cossu, M. Deka, A. Gandini, V. Luise, and G. Anselm, “Commons Based Peer Production in the Information Economy,” P2PValue (2016).
 C. Cheshire, and J. Antin, “The Social Psychological Effects of Feedback on the Production of Internet Information Pools,” Journal of Computer-Mediated Communication, 13, 1 (2008).
 M. Fuster-Morell, J. L. Salcedo, and M. Berlinguer, “Debate About the Concept of Value in Commons-Based Peer Production,” Internet Science (2016).
 See https://guifi.net, last accessed on 30th June 2019.
 R. Baig, R. Roca, F. Freitag, and L. Navarro, “Guifi.net, a Crowdsourced Network Infrastructure Held in Common,” Computer Networks: The International Journal of Computer and Telecommunications Networking, 90 (2015).
 J. Sánchez, “Architecture for the Commons: Participatory Systems in the Age of Platforms,” Architectural Design, 89, 2 (2019).
 R. Baig, R. Roca, F. Freitag, and L. Navarro. “Guifi.net, a Crowdsourced Network Infrastructure Held in Common,” Computer Networks: The International Journal of Computer and Telecommunications Networking, 90 (2015).
 Guifi.net. 2019. “Node Statistics,” Node Statistics Guifi.net (2019).
 G. Hardin, “The Tragedy of the Commons. The Population Problem Has No Technical Solution; It Requires a Fundamental Extension in Morality,” Science 162, 3859 (1968), 1243–48.
 E. Ostrom, Governing the Commons: The Evolution of Institutions for Collective Action (Cambridge University Press, 1990).
 E. Ostrom, “Understanding Institutional Diversity” (2009); M. Cox, G. Arnold, and S. Villamayor Tomás, “A Review of Design Principles for Community-Based Natural Resource Management” (2010).
 E. Ostrom, Governing the Commons: The Evolution of Institutions for Collective Action (Cambridge University Press, 1990), 88–102.
 F. B. Viégas, M. Wattenberg, and M. M. McKeon, “The Hidden Order of Wikipedia,” Online Communities and Social Computing, OCSC'07: Proceedings of the 2nd international conference on Online communities and social computing (2007).
 D. Rozas, A. Tenorio-Fornés, S. Díaz-Molina, and S. Hassan, “When Ostrom Meets Blockchain: Exploring the Potentials of Blockchain for Commons Governance,” SSRN Electronic Journal (2018), 8–20.
 S. Huckle and M. White, “Socialism and the Blockchain.” Future Internet, 8, 4 (2016), 49.
 P. De Filippi, and S. Hassan, “Blockchain Technology as a Regulatory Technology: From Code Is Law to Law Is Code,” First Monday, 21, 12 (2016).
 D. Rozas, A. Tenorio-Fornés, S. Díaz-Molina, and S. Hassan, “When Ostrom Meets Blockchain: Exploring the Potentials of Blockchain for Commons Governance,” SSRN Electronic Journal (2018), 21–22.
 A. Pazaitis, P. De Filippi, and V. Kostakis, “Blockchain and Value Systems in the Sharing Economy: The Illustrative Case of Backfeed,” Technological Forecasting and Social Change, 125 (2017), 105–15.
 R. Baig, R. Roca, F. Freitag, and L. Navarro. “Guifi.net, a Crowdsourced Network Infrastructure Held in Common,” Computer Networks: The International Journal of Computer and Telecommunications Networking, 90 (2015).
 See Baig et al. (2015) for further details.
 J. Sánchez, “Architecture for the Commons: Participatory Systems in the Age of Platforms,” Architectural Design, 89, 2 (2019).
Mereology is a formal concept which enters architecture as an additional formal category. Form is a rather ambiguous concept in architecture. So in this essay, first an investigation is conducted by contrasting two closely related concepts: shape and content.
Hans Trusack criticises the problem of shape for its shallow formalism and historical-theoretical indifference as a defensive strategy that evades the disciplines and difficulties of past and future. The distinction between the terms “form” and “shape”, following Tursack’s argument, is a “matter of generative process”. Both terms point to the production of visual expression. Yet while shape refers to the appearance of an object, form reflects the logic of transformation and operation within historical and theoretical contexts such as political and religious ideology, economics and technological background. Tursack criticised the strategy of shape in architecture, stating its lack of reference, it being “plainly, and painfully, evident”, and incapable of moving forward. Whereas form is difficult, disciplinary and requires historical and theoretical study, and yet promises the future.
Form has the advantage of being able to deal with complex relations due to its deep and continuously evolving intervention with content. The term form derives from the Latin word forma, is understood as the combination of two Greek words: eidos, the conceptual form, and morphe, the physical form. The complexity of form can be attributed to these differentiated meanings, yet complexity is compatible with agencies and relations. This can emerge further by conducting a brief historical review.
Ancient Greek architecture pursues the ideality in mathematics and proportion. The efforts made by architects in designing the Parthenon provides evidence of this feature. These operations tried to approximate the physical shape of architecture to the “ideal” form. Form reflects the pursuit of ideality and perfection in this period.
For Gothic architecture, there were more concerns about structure, and matter was pushed to its maximum capability to build as tall as possible for religious appeal. Consequently, structures were designed to be rigid and lightweight, and solid walls were replaced by glass windows, while flying buttresses supported the main structure to grow even taller. Consequently, astonishing space and fascinating transparency emerged.
Modernism claims that “form follows function”, rejecting traditional architecture styles. The reality of matter and the logic of technology eschewed decorations, proportions, or any subjective distortion of matter. The emphasis on the term “function” illustrates an ideology of treating architecture as a machine. Each part is nothing more than a component that has a certain feature inside this machine, and redundant decorations and details are removed to deliver this idea clearly. Without distractions, space becomes evident.
In the shift to postmodernism, the uniformity and the lack of variety of modernist architectures were reacted against, and a great variety of approaches emerged to overcome the shortcomings of modernism. Parametricism, for instance, has been promoted by the thriving of digital technologies. Designers are capable of more complex formal production, and architectural elements have become variables that can be interdependently manipulated. In this formalism, rigidity, isolation, and separation are opposed, while softness, malleability, differentiation and continuity are praised.
From the examples above, form is the embodiment of the relations between architecture and its motive in specific historical scenarios, while for shape, only the results are accounted for – relations are ignored, and architecture is treated as isolated physical entities, incapable of producing new relations. Different methodologies of dealing with architectural form also imply the variation of ideology in compiling form with content.
Mereology – An Approach of Architectural Form
In recent philosophical texts, a third notion of form is brought forward. Contrary to a dialectic of form and content, here investigations deal with the resonance of parts: the description of objects by their ontological entanglement only. The writings of the philosopher Tristan Garcia are a strong example for such mereological considerations. In his treatise Form and Object: A Treatise on Things (2014), Garcia investigates the ontology of objects with two initial questions, “… what is everything compose of? … what do all things compose?” The first question interrogates the internal, the elementary component of everything. The second interrogates the external, the totality of everything. For Garcia, the form of a thing is “the absence of the thing, its opposite, its very condition,” form has two senses, the “beginning”, and the “end”, which never ends. Form begins when a thing ends, it begins with different forms; in the end, since it has “endless end”, form ultimately merges into one, which is “the world”. Garcia defines an object as “a thing limited by other things and conditioned by one or several things.” The form of an object depends on what comprehends or limits this object. Every object is “embedded in a membership relation with one or several things”, they can be divided by defining limits, which is also a thing distinguishing one thing from another. Garcia’s argument adapts the concept of mereology. Form has two extremes, one toward the fundamental element of matter, and the other toward the world, comprehending everything. All things can always be divided into an infinite number of parts, and they can always be parts of another thing. Identifying parts or wholes within a section we can operate on can establish a limit. The relevance between form and mereology opens a new opportunity to inspect architectural form from a different point of view.
One of the first discussions about parts and wholes in modern philosophy was posed by Edmund Husserl, in Logical Investigation (1st ed. 1900-1901, 2nd ed, 1913), but the term “mereology” has not been put forward until Stanisław Leśniewski used it in 1927 from the Greek work méros (parts). Mereology is considered as an alternative to set theory. A crucial distinction lies between mereology and set theory in that set theory concerns the relations between a class and its elements, while mereology describes the relations between entities. The mathematical axioms of mereology will be used as the fundamental theory of developing the method of analysing architectural form.
Following Roberto Casati and Achim Varzi, the four fundamental mathematical formularisations of mereology are: “Relations are reflexive, antisymmetric and transitive. (…) First, everything is part of itself. Second, two different objects cannot be part of each other. Third, each part of a part of a whole is also part of that whole. Fourth, an object can be a part of another object, if both exist.”
Mereology can be a promising approach also for the reading of architectural form, as it emphasises relationships without reducing buildings to their appearance or function. However, such philosophical descriptions consider wholes and parts as mostly abstract figures. Therefore, a supplement could be developed to properly categorise the mereological relations in the field of architecture. Having the relations between form and mereology addressed, methodologies can be developed to access the analysis of architectural form. Mereology as a specific methodology for architecture is quite new. One of the first introductions can be found in Daniel Koehler’s book The Mereological City: A Reading of the Works of Ludwig Hilberseimer (2016). Here, Koehler departs from the modern city, exemplified through the work of Ludwig Hilberseimer to illustrate mereological relations in the modernist city. From the room to the house to the city to the region, Hilberseimer canonically drew the city as a hierarchical, nested stack of cellular spaces. However, through the close reading of its mereological relations it becomes clear that political, economic or social conditions are entangled in a circular composition between the parts of the city. Recalling Garcia’s discourse, and resonating with Leon Battista Alberti’s thesis, Koehler shows that the cells in Hilberseimer’s modernist city are interlocked. A house becomes the whole for rooms; a city becomes the whole for houses. By considering the city and its individual buildings equally, “the whole is a part for the part as a whole.”
Architectural Relations Between Parts and Wholes
Parts are not only grouped, packed and nested through different scales, but also in different relations. Specific relationships have been developed in different architectural epochs and styles. Mathematically, four general classes of relations can be drawn: whole-to-whole, part-to-part, whole-to-parts and parts-to-whole, while more specific subclasses can be discovered from each.
According to the mathematical definition, between wholes there exist complex relations, the whole could exist on any mereological level, and the complexity of relations between multiple levels are also accounted for. Whole-to-whole relations can become complex when considering multi-layer interaction, and more relations can be identified: juxtapose, overlap, contain, undercrossing, transitivity, partition, trans-boundary, intact juxtapose, compromised juxtapose.
A first glance of New York, gives the impression that it is quite heterogeneous, but underneath there is a city grid underlying the heterogeneity, and while the relations displayed in the grid are rather simple, all wholes juxtapose with one another. In comparison, in Siena, an Italian city, the urban space is quite complex, where boundaries of all wholes negotiate with others, the gaps in between are carefully treated, the nesting relations are extremely rich, and multiple relations from the diagram above can be found.
The whole-to-parts relation studies what the whole does to its part, namely in terms of top-down rules. The mathematical definition does not involve specific situations that a whole-part condition holds. Distinctions within individual contexts make a significant difference in clarifying an explicit relation. The situations for the whole can generally be classified into following types: fuse, fit and combine.
One of Zaha Hadid’s projects, Heydar Aliyev Centre, indicates the fusing relation. Architecture is represented as a smooth, fluid volume. The distinction between elements disappears, and this dominating power even extends to the external landscape. In order to maintain a continuous whole, parts are fabricated into a particular shape, having their unique unchangeable locations. The continuous whole excessively overwhelms the parts, yet not all parts are reshaped to fuse into the whole, and because the parts are small enough in relationship to the whole, the control from the whole is weakened, and parts are fit into the whole.
The third type is combining. An example for this relation is Palladio’s project Villa Rotonda. In this case, parts are obvious. The whole is a composition of the parts’ identities. However, the whole also holds a strong framework, in a rigorous geometric rule that decides positions and characters of parts. The arrangement of parts is the embodiment of this framework.
The parts-to-whole relation studies what the parts do to the whole, or the power of bottom-up relationships. The different situations of parts are also key parameters in validating a given relation. The classification of situations for parts are as follows: frame, intrinsic frame, extrinsic frame, bounded alliance, unbounded alliance.
Emil Kaufmann thoroughly investigated the innovative works by Claude Nicholas Ledoux in Three Revolutionary Architects: Boullee, Ledoux and Lequeu (1952). According to Kaufmann’s study, Ledoux’s works developed new compositional relations of elements from the Baroque. The characteristics of parts in Baroque architecture are rich, but tend to regulate the identities of all the elementary parts and fuse them together to serve the harmony of the whole, presenting the intrinsic framing. Ledoux’s work is an extrinsic framing, where the parts are relatively independent, with each element maintaining its own properties, and while consisting of the whole, they can be replaced with other identical components.
One of my projects in discrete aggregation of elements presents an unbounded alliance relation. The aggregation as a whole shows a form that is discretised (Figure 12), and does not pass any top-down instructions to its parts.
Part-to-Part Without Whole – The Ultimate Parts
For part-to-part relations, local interactions are emphasised, and interactions occur at multiple levels of compositions, where the part-to-part relations in some cases are similar to that between wholes. It has following classifications: juxtapose, interrelate, contain, partition, overlap, trans-juxtapose, over-juxtapose, over-partition, over-overlap.
Architects have been working on the possibility of removing the whole by studying the part-to-part relations. Several approaches have been developed, mainly through computation. Neil Leach considers the city as a “swarm intelligence”, bringing forward the potential of developing urban form with computational method. Leach encourages swarm intelligence for the interactions between agents (parts), which “offers behavioral translations of topology and geometry”, while fractals, L-systems or cellular automata are all constrained by some limitation. However, although swarm intelligence is based on the interaction of individual agents, it is always treated as a whole; all cells of CA are fixed in the background grid, which is also a whole. For fractals and L-systems, they can be subdivided into infinite parts, a transcendent whole where all parts grown from still exist. In the mereological sense, none of these cases can escape the shadow of the whole – strictly speaking, they are part-to-whole relations. To discuss the part-to-part relation in more depth, more investigation is needed to clarify the concept of part.
In The Democracy of Objects (2011), Levi Bryant claims that objects constitute a larger object by establishing relations with others, but this doesn’t alter the existence of objects, as he says, “all objects equally exist, but not all objects exist equally.” In Bryant’s discourse, this independence suggests the dissolution of the whole. Bryant proposes a concept of “regimes of attraction”, that includes the “endo-relation” and the “exo-relation”. The endo-relation indicates that the proper being of an object consists of its powers or what an object can do”, not the “qualities” emerging within an exo-relation. An object possesses “volcanic powers”, the stabilisation of the regime of attraction actualises it into a specific state. The concept of the whole reduces objects to this state, which displays only a section of their proper beings. The concept of regimes of attraction is against this reduction.
The regime of attraction can be linked to the notion of “assemblage” from Manuel DeLanda, however, there is a distinction between the two. Assemblage holds only the relation of exteriority, whereas regime of attraction maintains both relations of interiority and exteriority. In Assemblage Theory (2016), DeLanda reassembled the concept “assemblage”, which was originated from the French agencement. Created by Gilles Deleuze and Félix Guattari, this original term refers to the following meanings: the “action of matching or fitting together a set of components” – the process, and the “result of such an action” – the product.
DeLanda emphasised two aspects, heterogeneity and relations. As he indicated, the “contrast between filiations and alliances” can be described in other words as intrinsic and extrinsic relations.
The nature of these relations has different influences on the components. The intrinsic relation tends to define the identities of all the parts and fix them into exact location, while the extrinsic relation connects the parts in exteriority – without interfering with their identities. DeLanda summarised four characteristics of assemblage: 1) individuality, an assemblage is an individual entity, despite different scale or different number of components; 2) heterogeneity, components of an assemblage are always heterogeneous; 3) composable, assemblages can be composed into another assemblage; 4) bilateral-interactivity, an assemblage emerges from parts interactions, it also passes influences on parts.
DeLanda then moved on to the two parameters of assemblage. The first parameter is directed toward the whole, the “degree of territorialisation and deterritorialisation”, meaning how much the whole “homogenises” its component parts. The second parameter is directed toward the parts, the “degree of coding and decoding”, meaning how much the identities of parts are fixed by the rules of the whole. The concept of assemblage provides us a new lens of investigating these mereological relations. With this model, the heterogeneities and particularity of parts are fully respected. The wholes become immanent, individual entities, existing “alongside the parts in the same ontological plane”, while parts in a whole are included in the whole but not belonging to it, and according to Bryant’s discourse, the absence of belonging dispelled the existence of the whole.
From the study of regime of attraction and assemblage, this essay proposes a new concept – “the ultimate parts” – in which a proper “part-to-part without whole” is embedded. A part (P) horizontally interacts with its neighbouring parts (Pn), with parts of neighbouring parts (Pnp), as well as interacting downwardly with parts that compose it (Pp) and upwardly with wholes it is constituting which are also parts (Pw). This concept significantly increases the initiatives of parts and decreases the limitations and reductions of them. It doesn’t deny the utilities of the whole, but considers the whole as another independent entity, another part. It’s neither top-down, nor bottom-up, but projects all relations from a hierarchical structure to a comprehensive flattened structure. The ultimate parts concept provides a new perspective for observing relations between objects from a higher dimension.
One application of this concept is TARSS (Tensegrity Adaptive Robotic Structure System), my research project in MArch Architectural Design in B-Pro at The Bartlett School of Architecture in 2017–2018. This project utilises the features of tensegrity structures of rigidity, flexibility and lightweight. The difference is that rather than fixing parts into a static posture and eliminating their movements, the project contrarily tries to increase the freedom of parts as much as possible. The tensile elements have the ability to adjust their lengths collaboratively to change the general shape of the aggregation. Reinforcement learning is employed to empower the parts with objective awareness. The training sessions were set up toward multiple objectives that are related to architectural concerns, including pathfinding, transformation, balance-keeping, self-assembling and structural load distributing. This approach brings obvious benefits, as architecture design in this sense is not only about an eventual result, but about the dynamic process of constantly responding to the environmental, spatial or functional requirements. The premise is to treat parts as ultimate parts whilst retaining their objectivity and being able to actively interact at all mereological levels without limitations.
The concept of ultimate parts brings forward a new relation of “part-to-part without whole”. This new relation belongs to a higher dimension. The details and essence of objects are simultaneously displayed, without being obscured by the compositional structure. Analogised with spatial dimensions, a 3-dimensional cube simultaneously shows all its faces and interior in 4-dimensional space. The significance is that it opens vast new perspectives and operational methodologies in the architectural design realm. Especially with the advancement in robotics and artificial intelligence, this type of new relationship enables greater opportunities by regarding machines as characters with immense potential to work with us, instead of for us. The role of designers would be very much like “breeders of virtual forms”, who do not rule the form, but guide it towards the demands. This moves away from anthropocentric design by overcoming part-to-whole with part-to-part.
 H. Tursack, "The Problem With Shape", Log 41 (New York: Anyone Corporation, 2017), 53.
 Ibid, 50.
 L. Sullivan, "The Tall Office Building Artistically Considered", Lippincott's Magazine (1896), 403–409.
 T. Garcia, M. A. Ohm and J. Cogburn, Form And Object (Edinburgh: Edinburgh University Press, 2014), 19.
 Ibid, 48.
 Ibid, 77-78.
 Ibid, 145.
 E. Husserl, Logical Investigation (London: Routledge & K. Paul, 1970).
 Stanisław Leśniewski, O podstawach matematyki [trans. On the Foundations of Mathematics], I-V, 1927-1930, Przegląd Filozoficzny, 30 (1927), 164–206; 31 (1928), 261–291; 32 (1929), 60–101; 33 (1930), 77–105; 34 (1931), 142–170.
 R. Casati and A. C. Varzi, Parts and Places: The Structures of Spatial Representation (Cambridge, Massachusetts: MIT Press, 1999).
 L. Hilberseimer, The New City: Principles of Planning (P. Theobald, 1944), 74-75.
 D. Koehler, The Mereological City: A Reading of the Works of Ludwig Hilberseimer (Transcript, Verlag, 2016), 182.
 E. Kaufmann, Three Revolutionary Architects, Boullée, Ledoux, And Lequeu (Philadelphia: The American Philosophical Society, 1968).
 N. Leach, "Swarm Urbanism", Architectural Design, 79, 4 (2009), 56-63.
 L. Bryant, The Democracy Of Objects (Open Humanities Press, 2011), 290.
 M. DeLanda, Assemblage Theory (Edinburgh: Edinburgh University Press, 2016), 2.
 Ibid, 19-21.
 Ibid, 12.
 L. Bryant, The Democracy Of Objects (Open Humanities Press, 2011), 273.
 M. DeLanda, "Deleuze And The Use Of The Genetic Algorithm In Architecture" (2001), 3.
Object-oriented programming in blockchain has been a catalyst for philosophical research on the way blocks and their nesting are perceived. While attempting a deeper investigation on the composition of blocks, as well as the environment that they are able to create, concepts like Jakob von Uexkull’s “Umwelt” and Timothy Morton’s “Hyperobject” can be synthesised into a new term; the “Hyperumwelt”. The Hyperumwelt is an object that is capable of creating its own environment. By upscaling this definition of the Hyperumwelt, this essay describes objects with unique and strong compositional characteristics that act as closed black boxes and are able to create large scale effects through their distribution. Hyperobjects are able to create their own Umwelt, however when they are nested and chained in big aggregations, the result is a new and unexpected environment: the Hyperumwelt.
In his book Umwelt und die Innenwelt der Tiere (1921) Uexkull introduced the notion of subjective environments. With the term “Umwelt” Uexkull defined a new perspective for the contextualisation of experiences, where each individual organism perceives surrounding elements with their senses and reinterprets them into its own “Umwelt”, producing different results. An Umwelt requires two components: an individual and its abstracted perception of its surroundings. Based on this process and parameters, notions of parthood and wholeness in spatial environments, and the relations that they produce with interacting elements, become relevant.
Space as a Social Construction
For Bill Hillier and Julienne Hanson these two parameters related to society and space, writing that “society can only have lawful relations to space if society already possesses its own intrinsic spatial dimension; and likewise space can only be lawfully related to society if it can carry those social dimensions in its very form.” What Hillier and Hanson argue is that the relation between the formation of society and the space is created by the interaction between differing social environments. Hillier and Hanson essentially make use of a mereological definition of the environment that states that parts are independent of their whole, the way that society is independent from its space, but at the same time societies contain definitions of space. Space is therefore a deeply social construction.
As Hillier and Hanson outline, our understandings of space are revealed in the relations between “social structure” and “spatial structure”, or how society and space are shaped under the influence of each other. Space is a field of communication. Within a network of continuously exchanged information, space can be altered as it interacts with the people in it. However, this approach can only produce limited results as it creates environments shaped by only two parameters, humans and space. At this point is where Hillier and Hanson’s theory fails, as this way of understanding the environment relies only on additive information produced by interactions. If we were to expand this theory into the kind of autonomous learning mechanism that is mandatory for processing today’s computational complexity, we would end up with a slow, repetitive operation between these two components.
Hyperobjects to Hyperumwelt
Another perspective that is elusive from Hillier and Hanson’s understanding of the environment is how social behaviour is shaped by spatial parameters. Timothy Morton’s object-oriented ontological theory contradicts this anthropocentric understanding of the world. In The Ecological Thought (2010) Morton presents the idea that not only do we produce the environment but we are also a product of it. This means that the creation of things is not solely a human act non-human objects cannot partake in, but rather an inherent feature of any existing object. For Morton, complexity is not only a component of society and space, but extends complexity to an environment that has objects as its centre and thus cannot be completely understood. He calls these entities ‘Hyperobjects”.
While Morton uses the term Hyperobject to describe objects, either tangible or intangible, that are “massively distributed in time and space as to transcend spatiotemporal specificity”. The term can be reinterpreted to describe an environment, rather than an object, which is neither understandable nor manageable. This environment – a Hyperumwelt – is the environment constructed by Hyperobjects. A Hyperumwelt is beyond comprehension due to its complexity.
The term Hyperobject is insufficient as it retains its own wholeness. This means that all components inside a Hyperobject cannot be seen (as it acts like a black box of information) but can only be estimated. Morton described the Hyperobject as a whole without edges. This stems from Morton’s point of perception, as he puts himself inside of the object. This position makes him unable to see its wholeness and thus it leaves him adrift of its impact, unable to grasp control of it. Here, also, the discussion opens about authorship inside the environments and what Morton suggests is that Hyperobjects have their own authority and there is nothing that can alter them or specify their impact on the environment.
A Tree in a Forest
Yet there is also no need for the Hyperobjects to be clearly understandable. In terms of the Hyperumwelt, Hyperobjects can remain vast and uncomprehended. What is now needed are the implications of distributing nested Hyperobjects, seen as black boxes, inside an environment. An Umwelt is an environment constantly altered by the perceived information. This makes the Hyperumwelt a whole with porous edges that allows the distribution, and the addition or subtraction, of information. Another difference is the external position that the Hyperumwelt is perceived from, meaning that there is no need for it to be part of the environment. Since what is important is the distribution of the objects within the Hyperumwelt, a distant point of view is needed in order to detect the patterning of the distributed objects. While it will remain difficult to decipher and discretise the components, the patterns that are created can be seen.
While the Hyperobject is a closed whole of parts that cannot be altered, a Hyperumwelt is an open whole of wholes that uses objects as its parts. So, while the Hyperobject gives us no authority over its consequences, the Hyperumwelt bypasses this in order for its wholeness to be controlled. Yet what is important for the Hyperumwelt is not the impact of one object, but the impact of multiple objects within the environment. This synthesis and merging of objects and their relations produces a new reality which may or may not be close to the reality of the single objects. A Hyperobject is looking at a black box – say, a tree – and knowing there is a pattern – such as a forest – and a Hyperumwelt is looking at the tree and knowing the impact that it has on the forest and the impact that the forest creates in the environment.
 J. von Uexküll, Umwelt und Innenwelt der Tiere (Berlin: J. Springer, 1909), 13-200.
 T. Morton, Hyperobjects: Philosophy and Ecology After the End of the World (Minneapolis, Minnesota: University of Minnesota Press, 2013).
 J. von Uexküll, Umwelt und Innenwelt der Tiere (Berlin: J. Springer, 1909), 13-200.
 B. Hillier and J. Hanson, The Social Logic of Space (London: Cambridge University Press, 1984), 26.
 T. Morton, The Ecological Thought (Cambridge, Massachusetts: Harvard University Press, 2010).
 Ibid, 110.
 T. Morton, Hyperobjects: Philosophy and Ecology After the End of the World (Minneapolos, Minnesota: University of Minnesota Press, 2013).
 T. Morton, Being Ecological (Penguin Books Limited, 2018).
This interview took place on April 11th, 2017 at the office of Herman Hertzberger in Amsterdam, with questions by Daniel Koehler.
Daniel Koehler: After all your years as a teacher, maybe it would be a good departure for conversation if you can tell us what is your favourite exercise for teaching architecture?
Herman Hertzberger: Well, my favourite exercise is making a housing environment, where small children could live and play outside. This is an old-fashioned thing, but I am absolutely convinced that children should play in the streets in order to find out about the world and to learn about the good and the bad things that exist. I am afraid that today in urbanism you find high-rises, and the immense distance from the living unit to the street is a problem. Consequently, cities only have playgrounds with fences around, and there children are safe to play. But the world is not only about safety, the world is about finding out how far you can go in your life.
Can you tell us a little bit about how you began to communicate as a group during the beginning of structuralism?
We started a school! We had seminars for discussion, where everyone brought in something to discuss. And then we had the Forum editorial staff. There was Aldo van Eyck, Jacob Bakema and others who are less well known (Dick Apon, Joop Hardy, Jurriaan Schrofer and Gert Boon). There was an enormous amount of communication. Every Tuesday night we saw each other with no exception, where we discussed the next issue of the magazine.
And, what was the relevant media at the time you started to develop your ideas?
Magazines were very important. We used to have three or four Dutch magazines, two French, two English, one American, two Swiss, two Italian. They were all on the table. “Did you see that building? I think it is good.” And then we had a discussion. Today we still get some magazines, but today you get all the information from around the world in one click. That is fantastic, the possibilities today are just immense. It is more information, a lot of pieces of an enormous puzzle. But is it also possible to put it together? I hope the younger generation can.
The magazine Forum was for me a sort of postgraduate study. At that moment I started to see the work, the hands and eyes of real architects. That helped me to start thinking. And there were connections to many other architects from all over. There were conferences, and there you saw people. At Delft University, where I was teaching, we invited all the architects we were interested in. We are still doing this.
When one reads the current literature, one can draw two different issues. On the one hand a common critique of functionalism in architecture, and on the other hand, the influence of new ideas coming from sociology. Would you say that this enormous explosion of ideas and diversity of projects was a response to architectural problems or were these new concepts coming from other fields prescriptive to your projects?
First of all, there is nothing coming from sociology. I have little or no connection to sociology. Sociology is the science of human relations. But you do not need to go into this science as an architect. Architecture is a matter of using your eyes and ears to look into the world and see what needs to be done. But today, architecture is driven by algorithms and rules. All the rules, saying you should do this or that, you are supposed to not do this. Architecture is then reduced to problem-solving. You must be aware of that mistake. Architecture is not problem-solving. Of course, you have to solve problems, but this is only one aspect of architecture. It starts to be architecture when it provides more freedom to people, opening the possibility that things are getting better than they were before.
Can you give an example?
A dwelling needs to have a balcony. Why? To let people go outside, and there are rules about the size of your balcony. Most architects think: well, I included a balcony. But they should base the form and dimensions of a balcony on the needs of daily life. Such as sitting in a corner without being seen by others or not being disturbed while reading your book yet with the possibility of having contact with your neighbour. Second, you may want to be able to have a meal with your family. You maybe want to have flowers and plants. In a way, this is culture. Make that list, and when you design a balcony, be sure that all the points you have listed have also been fulfilled. In this way you increase people’s freedom. Most balconies do not do that. On most balconies all you can do is sit. Most architects don’t think, they don’t look at what is going on. And then, of course, the developer says, “It should not cost more, so we have to make it small.” So I have rules independent of the developers. For me, it starts to become good when those rules are going to be met. This method works for every part of the building, from a dwelling, to a living unit, to a street, to the school. In a school, you can design where the black board in a classroom is going to be. And you have to think about what a school might be. I don’t need sociologists for that. Sometimes, sociologists can tell you some interesting things, but you have to think, and in the first place, look for yourself.
Your communal spaces are famous for their human scale, like the doorstep. I think that this down-scaling of the city to elements of a building enables you to design the building as an open system. For me, it seems you draw a difference in creating a building as a building and designing a building as a city.
For me, city, architecture, and building are very much related. Aldo van Eyck believed that making architecture is always making things more inside than before. Aldo van Eyck said, “Whatever you do, it is supposed to always increase the inside quality.” When you want to go outside, you go to the fields. There you have the horizon, you have the clouds and the openness. A city is for exchange – exchange of goods, of ideas. Cities are mostly based on trade, and on having a cinema, having shops, having communal things, being together.
Aldo van Eyck also claimed that the city should be a big house. I think that is a dangerous thing to say because the city is not the house where you are yourself, or where you are enclosed. The city should never be enclosed but always open, in connection with the whole world. It is the place where you see the airplanes flying above you. But it is an inside space in relation to the open field. And a building is, in fact, a small city. Make a building as a small city to have the emphasis lay on communication and exchange.
But most buildings are private territories with public corridors. How narrow can a public corridor be? It cannot be a centimetre larger, because this would cost money. Means of communication are considered extra. You can sell the dwellings but not the corridors. As a result, most buildings have very beautiful apartments and very small corridors. I am pleading for buildings where the corridors are streets. I try to put more emphasis on the communal spaces in a building.
When you consider a building as an open system, what role does the boundary between inside and outside have? Do you think that these open systems have an outside or do you think of them as endless? What is their relationship to the context and environment?
City is not just buildings but the space in between the buildings as well. The edge of building is forming the space of the city. You have to conceive of the edge of a building not as an end where the outside starts. You must see it the other way around, as a wall in the interior space of the city. The idea of the building as city is to put buildings in such a relation that the space in-between them is as important. This is something that is completely lost. It is also considered nostalgic. But look to New York. In New York, you have these high skyscrapers, but you also have very nice streets. When I am in Manhattan, I feel quite enclosed. That is because of the very strict system of the grid, and the building lines by which the streets are defined, and the blocks in between are open.
In one of your articles [Open City, 2011], you rightly point out that most of today’s housing projects consciously exclude communal spaces, and focus only on the assembly of private areas without any spatial linkages between them. Private areas are protected to one another rather than connected. A common – and I think dangerous – justification for such a design refers to changed economic circumstances, and most cynically, to the death of the welfare state. Would you say architecture is so dependent on economics?
Every square meter is supposed to generate a fee, so public space will be reduced to a minimum. Architecture has become business. And that makes the position of the architects to contribute to better spaces and towns very difficult.
But then architects are even more important.
Important as long as you are able to be aware of what sort of culture you are living in. I cannot give you the answer what to do. You have to explain and fight. But you need clients who believe in the architect. Things are very materialistic today. But there are also very interesting initiatives. For example, in Rotterdam, you have these old industrial halls which could be reused without high costs. Add a little paint, and it works. There are ways today that are contradicting this idea of architecture as economics. There is a lot for you to invent.
When I told a friend that we will visit you as one of the main protagonists of structuralism, his response was: ‘Wait a moment! Herman Hertzberger is not a structuralist; he is a humanist.’
Can you not be a structuralist and a humanist at the same time? Is this contradictory?
I think what my friend was pointing at is that there is a difference between structuralism as a style and structuralism as modus operandi, as a form of organisation and composition.
Style has to do with aesthetics, but aesthetics is a pitfall. Most architects think making something beautiful is all that architecture is about. But you can’t make something beautiful, it is impossible. That doesn’t work. What you can do is make a painting which is striking, and shows you something you never saw before that makes you happy or fall in love with the painting, and then we decide this is a beautiful painting. But in architecture, don’t spend energy on trying to make something beautiful. Make it work. Then you may hope that someone says this is beautiful. For instance, the composer of music Arnold Schönberg said, “Do not do what others consider beautiful, but just what is necessary to you.” I like a building because it works. When someone if I think it is beautiful, then I say, when you are in love it is going to be beautiful. Beauty comes as a result. But you cannot say, now I am going to make it beautiful. Beauty is a pitfall for architects.
Structuralism means there are simple rules that enlarge the amount of free space that you can achieve. I took the grid as an example earlier. The very rigid system of the grid allows you to be more free in the blocks in between. All of the blocks can be different; some high, some low. It is an enormous mosaic of possibilities that is held together by the grid. When you know what rules you have to use, you can be creative. It is a misunderstanding that the one contradicts the other.
It is interesting that you describe a rule as a form of enclosure, as a form of an inside.
If something is not limited you create chaos. Rules prevent you from chaos, and within rules you can be creative. Noam Chomsky [the linguist] uses the words competence and performance. The structure of language is its competence, it is its capacity to express. And performance is what you are actually expressing with it. In language you have grammar, but every individual talks in his or her personal way using the same rules.
Would you say that you have a grammar and vocabulary then? Do you have certain elements that you are frequently using? You were talking earlier about balconies and streets. In your work do you consider elements repeat structurally, which can re-emerge in different styles, but with similar performance? Or do you begin each project with a new grammar?
I do not use the same grammar for every building. I could, but I want to try different things. There are many people who thought housing should be produced in a factory, like cars. It is such a simple idea. But it doesn’t work, because every location has its own needs, whereas a car is the same everywhere. So, you can not use the same grammar. I use another grammar for a school and another for housing for instance. Some things have a similar grammar, like how you make a door, which works in most cases.
Do you have a particular vocabulary of elements that reappear during your career in different articulations and styles?
Architecture should accommodate people and things that people are concerned with. I use this everywhere. To give you a simple example: when I make a column, most of the time I design it with a base for people to be able sit on it. This is for me an accommodating device. It always works. This sort of thing is universal in my opinion: the idea of accommodation. Another example is the handrail of a stair. I always make a handrail that guides you where to go, making the end of it in such a way that even without looking you have the feeling that this is the end of the stairs. Everything I do tries to consider how it works for people. However the point is that it should be friendly to people, but not soft.
Friendly architecture! This is a wonderful conclusion. Thank you, Herman Hertzberger for sharing your time and thoughts with us.