Search Results
146 results found with an empty search
- In conversation with Paul Beuchat
By Renee Papaluca < Back to Issue 3 In conversation with Paul Beuchat By Renee Papaluca 10 September 2022 Edited by Zhiyou Low and Andrew Lim Illustrated by Ravon Chew Next Paul is currently a postdoctoral teaching fellow in the Faculty of Engineering and Information Technology. In his spare time, he enjoys overnight hikes, fixing bikes, and rock climbing. Note: The following exchange has been edited and condensed. What was the ‘lightbulb moment’ that prompted you to study science? I often say that I chose engineering a little bit by not wanting to choose anything else. I think it also played into my strengths back in high school. I wasn't particularly into English, history or languages but I really enjoyed physics, chemistry and maths. So, that already drew me to science broadly. What ended up directing me towards engineering, and particularly mechanical engineering, was just always tinkering at home. My dad was always tinkering and building things. We had a garage with all of the tools necessary, and I had free rein to pull things apart and put them back together. Mechanical engineering was a way of taking a more formal route of enjoyment into the hobby. Why did you choose to pursue a research pathway? After I finished my double degrees in Science and Engineering, I got a job, which I enjoyed. It was fun working with a bigger team. In this case, it was an oil and gas company with some pretty big equipment involved. This wasn’t just tinkering with something little in the garage, but something on an industrial scale. At some stage, though, I felt like there was a bit missing. There was a research arm as part of the company, but that wasn't somewhere that I could get to. I was excited by the kind of work being done in that area, and I saw a PhD as a way of pursuing that love so that I could then work on those sorts of exciting things. What advice would you give to students considering a research pathway? Certainly, while I was a PhD, all the postdocs would say that the PhD was the best time of their life. Then the PhDs would say that the Masters was the best. So, be prepared for it to be hard. The advice is to be passionate about the topic and not be fearful about uncertainty or knowing the exact topic straightaway. Also, you likely will need a lot of support to get through the hard parts. It’s nice to have tangential input in the form of seminars, visiting academics from other institutions or even from PhDs in the same group or department. This input gives you new knowledge, new exciting fields and new industry connections. What sparked your love of teaching? My original intention was to complete my PhD, gain the relevant skills and return to the industry. My passion for teaching was sparked during my PhD experience; I got to supervise Masters students that are working on a larger project with me. It was a close collaboration with someone, where you start the process of teaching them whatever the topic is. You work on it together, and eventually, the student becomes the master. They can now guide you along, as well as having vibrant discussions together. That's what I find exciting about tertiary education more broadly - we all are pushing the limits of engineering to achieve better outcomes together. What does your day-to-day life as a teaching fellow look like? One of the focuses of my position was to include more project-based teaching, i.e. to include more hands-on education and work in the classroom, which was not included previously. I got the opportunity to create a new subject. I initially spent a lot of time developing what it was going to be. My day-to-day work included choosing new topics to add to the subject and linking them to a hands-on project, like a ground robot. There's a whole bunch of work that goes into designing a robot and the relevant software on top of preparing lecture slides and delivery—all these bits and pieces that make up a subject. Scattered throughout all this is teaching research; the teaching team assesses the students, and I need to assess the teaching itself. For instance, I need to understand what is being attempted in a particular class, what we are intending to achieve and how this aligns with the current best practices in education research publications. What advice would you give to students considering academic teaching as a career? One of the very nice things here at the University of Melbourne is the support teaching staff can receive through the Graduate Certificate of University Teaching. This gives you insight into and guidance on how to tackle the whole field. For instance, one of the lecturers mentioned that you have to be passionate about teaching because it has its ups and downs. Certainly, while developing a new subject, I found it to be quite stressful. It’s a different way of thinking, and all-new terminology, which is exciting and scary, and that took me a little bit by surprise. Where I shot myself in the foot the most was trying to do too much. I was in a very lucky position where I had free rein to make a subject as hands-on as possible, which opened the floodgates to possibilities. Prioritising was extremely important. It's not that you don’t try everything, but trying too many new exciting ideas at the same time means they probably are all going to fail or take an exorbitant amount of time to implement properly. Being realistic in my instruction was important. Also, having a mentor or someone you can talk very openly with was helpful. What are your future plans? For now, my intention is to stay in teaching. I’d like to push this position to the limits of what I can achieve and see where it takes me. I can also imagine the level of curriculum redesign in shifting whole courses to project-based learning. Current reports, like from the Council of Engineering Deans, are pushing for all engineering education to shift over to project-based learning within the next five to ten years. I’d like to continue teaching, with a view to contributing to higher-level curriculum development. Previous article Next article alien back to
- Entwined: A Hug Story | OmniSci Magazine
< Back to Issue 9 Entwined: A Hug Story by Elise Volpato 28 October 2025 Illustrated by Esme MacGillivray Edited by Steph Liang Ranging from Will’s heartbreaking collapse in Sean’s arms (Good Will Hunting (1)), to Sheeta and Pazu’s cheerful embrace (Castle in the Sky (2)), to Love Actually’s opening scene (3), hugs are everywhere. In cinema, songs, poems or artworks, they embody strong emotional connections. A s we observe and experience affectionate physical touch in various contexts, let us not forget about the importance of emotional connections in our own lives . Sharing a hug with your lover(s), your friends, your family, your pets; it seems to be an ordinary action… for extraordinary benefits. When hugging, we can all feel pleasant emotions such as serenity, joy, love. But what is the science behind being entwined to someone? Both psychologists and neuroscientists have puzzled over this question, and proposed potential explanations from numerous studies. Before we dig deeper into the warm world of hugs, I invite you to take some time to reflect on your own experiences: is physical contact important for you? What makes a good hug? Does being entwined to someone mean something to you? We will see that the perspectives on hugging differ through culture, physiology and psychology. Let’s now unknot the strings of our health through the lens of hugging! Hugging as a cultural practice Hugging is embedded in culture. It is often considered as a social greeting, either at the moment of an encounter between two people, or when they say goodbye to each other. Hugging, rather than handshaking, implies a reduction of interpersonal distance, greater emotional involvement and the willingness to show it. It is important that both people want this closer contact, as physical proximity is not appreciated by everybody. This is where particular cultural customs will feel natural for some and uncomfortable for others, depending on the greeting expectations and the person’s disposition to comply with them. Certain cultures will favour handshakes, kisses on the cheeks, a quick tap on the shoulder, or head nods (4). Hugging is not a universal practice. In fact, hugs are more common in warmer countries (alongside other forms of social touch), and within young people and females, but less practiced by conservative and religious populations (5). Physical touch seems less prevalent in Asian cultures – for instance, compared to countries such as Mexico, Costa Rica, or Sweden, China often has the lowest levels of hugging, whether between partners, friends, or a parent and their child (5). Hugs are also a symbol of cohesion, with sports teams’ group hugs providing motivation before a match or celebration after the victory. Interestingly, most studies into this have been conducted in Europe and Northern America, reflecting a bias in the cultural significance of hugging and what we take it to symbolise. Cultural context highlights that hugging serves multiple functions: greeting, social support, but also group cohesion and strengthening relationships. Why your body wants a hug Whether the cultural environment promotes hugging or not, this action inevitably has a physiological impact on people. A primary belief is that the physical warmth of an embrace makes the body feel relaxed, comfortable, and protected. It does not stop there, with hugging triggering various biochemical and physiological reactions, such as a higher magnitude of plasma oxytocin (bonding hormone), decrease in cortisol (stress hormone), and lower blood pressure (6). Hugging also reduces colds, promoting a more efficient immune system, and daily hugging predicts lower levels of two proinflammatory cytokines (7). Clinically, inflammation is a significant health marker, and plays a role in both mental and physical diseases. These results support the “affection exchange theory”, stating that affectionate interpersonal behaviour decreases stress and enhances immunity (excluding mitigating factors). Interestingly, studies show a general preference for right-arm given hugs. This effect is bigger (92%) when there is little emotional connection between huggers; for instance, in a “Social Media Challenge” setting where one person has their eyes covered and is hugged by random people (8). On the other hand, only 59% of people in international airport arrival halls (who are likely strongly connected to each other) hug with the right arm (9). These findings align with the “right hemisphere theory”, which states that the right hemisphere of the brain is dominant in emotional processing. Therefore, in situations of emotional hugging, the right hemisphere (which controls the left side of the body) takes the lead, so individuals hug each other with their left arm. Hence, emotional networks in the brain affect our hugging behaviour. Mind and perception If physical health can be bettered by regular hugs, we should not forget the undeniable links between physiology and mental health. Indeed, they are entwined in a virtuous circle. Due to decreased blood pressure and pulse, stress regulation is enhanced. This regulation is essential to emotional stability, for example before public speaking (10). Cortisol levels – which are related to both physical and psychological stresses - are lowered following a twenty-second hug, compared to no physical connection. This “well-being hack” works either with another person or even by self-hugging (11). Furthermore, research suggests that oxytocin has analgesic effects and influences pain processing areas in the brain (12). Pain is often thought of as a physical process, but it is multifactorial. In psychology, the “gate-control theory” (13) explains that a “gate” in the spinal cord exerts effects on pain perception by combining excitatory inputs from noxious stimuli with inhibitory ones. Thus, pain perception is modulated by both physical, ascending factors, and psychological, descending elements. As oxytocin release aids pain management, human psychology is positively influenced by the benefits of this neuromodulator, as well as the conscious, pleasant perception of hugging. Clearly, our mental health is particularly impacted by physical connection. As there is a lot of individual variability in the way people enjoy embraces, we may wonder whether hugs are more context-dependent or trait-dependent. When we look at personality traits, extraverted individuals tend to take the initiative in hugging, illustrating their spontaneity and warmth. On the other hand, neuroticism shows a tendency to social withdrawal combined with low self-esteem (14). While personality traits can be present from birth, some elements depend on our experiences during infancy. This is particularly relevant for attachment styles. When elaborating on this theory in 1969, Bowlby (15) described how it was essential for a child to not only experience affectionate and encouraging language, but also caresses and physical embraces, in order to develop a secure attachment. Throughout our entire lifespan, regular and adequate physical touch is hugely beneficial to human development. Conclusion The science behind hugging reveals multiple benefits. As long as the embrace is agreed on by all parties, there are minimal negatives, and the hug makes way for social, physiological and psychological advantages. As human beings, we are a highly social species that craves social connection, whether it is through physical bonds, emotional links, or both (hint: a key factor to achieve both is hidden in this article). Being interlaced is a marvellous way to improve your day, and even your life – go increase your oxytocin levels, I promise it is worth it. In the end, feeling entwined tells a meaningful story: a hug-story. References Scalia P, ed. Good Will Hunting . Miramax Films; 1997. Seyama T, Kasahara Y, eds. Castle in the Sky . Toei; 1986. Moore N, ed. Love Actually . Universal Pictures; 2003. Ocklenburg S. The Psychology and Neuroscience of Hugging . Springer Nature Affective Interpersonal Touch in Close Relationships: A Cross-Cultural Perspective. ResearchGate . doi: 10.1177/0146167220988373 Grewen KM, Girdler SS, Amico J, Light KC. Effects of Partner Support on Resting Oxytocin, Cortisol, Norepinephrine, and Blood Pressure Before and After Warm Partner Contact. Psychosomatic Medicine . 2005;67(4):531-538. doi: 10.1097/01.psy.0000170341.88395.47 Lisa, Floyd K. Daily Hugging Predicts Lower Levels of Two Proinflammatory Cytokines. Western Journal of Communication . 2020;85(4):487-506. doi: 10.1080/10570314.2020.1850851 Packheiser J, Rook N, Dursun Z, et al. Embracing your emotions: affective state impacts lateralisation of human embraces. Psychological Research . 2018;83(1):26-36. doi: 10.1007/s00426-018-0985-8 Turnbull OH, Stein L, Lucas MD. Lateral Preferences in Adult Embracing: A Test of the “Hemispheric Asymmetry” Theory of Infant Cradling. The Journal of Genetic Psychology . 1995;156(1):17-21. doi: 10.1080/00221325.1995.9914802 Grewen KM, Anderson BJ, Girdler SS, Light KC. Warm Partner Contact Is Related to Lower Cardiovascular Reactivity. Behavioral Medicine . 2003;29(3):123-130. doi: 10.1080/08964280309596065 Dreisoerner A, Junker NM, Schlotz W, et al. Self-soothing touch and being hugged reduce cortisol responses to stress: A randomized controlled trial on stress, physical touch, and social identity. Comprehensive Psychoneuroendocrinology . 2021;8(100091):100091. doi: 10.1016/j.cpnec.2021.100091 1.Boll S, Almeida de Minas AC, Raftogianni A, Herpertz SC, Grinevich V. Oxytocin and Pain Perception: From Animal Models to Human Research. Neuroscience . 2018;387:149-161. doi: 10.1016/j.neuroscience.2017.09.041 Melzack R, Wall PD. Pain Mechanisms: A New Theory. Science . 1965;150(3699):971-978. Forsell LM, Åström JA. Meanings of Hugging: From Greeting Behavior to Touching Implications. Comprehensive Psychology . 2012;1:02.17.21.CP.1.13. doi: 10.2466/02.17.21.cp.1.13 Bowlby J. Attachment and Loss: Attachment .; 1969. Previous article Next article Entwined back to
- Glowing Limelight, Fashioned Stars | OmniSci Magazine
< Back to Issue 8 Glowing Limelight, Fashioned Stars by Aisyah Mohammad Sulhanuddin 3 June 2025 Edited by Kylie Wang Illustrated by Jessica Walton Good evening Rose Bowl, Pasadena! The crowd erupts into a roar, the stadium air overcome with a thunder of adulation. Between throngs of teenagers tearing through streets in pursuit of the Beatles, concert-goers fainting at the sight of Michael Jackson, and Top Tens of the day made to navigate flirty fan calls on daytime TV in front of live audiences (1), pop history as we know it has always revolved around the deep, fanatic reverence of the star . Stars in all corners of the entertainment cosmos, be it music, film or TV, have long had their lives glamorised. Tales told of luxurious jet-setting, post-show mischief and infamous public appearances peppered with paparazzi. Fame turned into fables, circulated eagerly by the wider populace. Having avidly followed a plethora of musicians, actors and comedians at different points of my own life, the gurgling vortex of stardom culture has remained ever-intriguing. Why do our relationships with stars mean so much to our society, and have they shifted over time? Public perceptions & parasocial relationships Our journey begins with the making of a star. A star is born from an assemblage of artistic choices: artwork, stage personas, press releases, bold onstage costumes and more, which constellate into a fashioned image. Or, a ‘manufactured personal reality’ (2). This reality is what audiences draw upon when forming attachments to stars, a process that moulds complex, contradicting human beings into idealised forms that may resonate, validate or provide meaning to them. The mid-century women empowered by the feminine sexuality and intelligence of Marilyn Monroe (2), or the working class Eastern European following of Depeche Mode who saw the band as an emblem of social rebellion under the USSR in the late 80s (3), are such examples. Such attachment gives rise to the infamous ‘parasocial relationship’ (PSR). An often derisive term aptly used today to call out toxic, boundary-crossing online fan behaviour, parasocial relationships at their core simply encompass socio-emotional connections formed with media figures (4). In it, audiences extend emotional energy, time or interest towards figures that whilst unreciprocated, create a perceived idea of intimacy similar to that of two-way relationships. For the audience, PSRs can evoke feelings of safety, trust and various forms of devotion, self-strengthened through personal habits – think dressing like a favourite ‘bias’, or diligently watching a favourite director’s closet picks. PSRs have historically been one-sided. Audience reactions to sensation and scandal have had the power to make or break an artist’s image, but restricted channels of dialogue meant that direct two-way feedback was often “fragmented” (2). The influencing power of the star’s image lay within reach of the star themselves, and more often than not, was shaped by the wider commercial agendas of their agency or labels. That is, until recently… The rise of the Internet Whilst the glitz and glamour of stardom remains strongly relevant, we can focus on the advent of the internet as the most powerful force in reshaping the relationship between fan and star. Termed the “o ne and a half sided” PSR (4), seen today is a shift in power dynamics towards one of increased fan-star symbiosis. As the theory notes, technology has allowed for greater perceived proximity and reciprocity, blurring the line between social and parasocial. Under the extensive nature of the current digital world, our internet presence has become increasingly considered a material extension of our real-life selves (4), whether through Zoom calls, real-time story updates or live vlogs. Direct messages or comments that allow instant reply have muddied the realm of physical and virtual reality, thus leading audiences to feel ‘physically’ closer to the figures in question. This decrease in constructed social distance has fostered notions of reciprocity, viewing stars as people they can reach out to and touch, converse with, and most importantly, influence in return – regardless of any actual ability to do so (4). As we witness stars defend their personal choices against an onslaught of ‘netizen’ backlash or wryly reply to a barrage of invasive thirst tweets (5), we see the increased power that global audiences have over said stars’ images. Eroded power barriers between the star and fan have heightened both positive and negative emotional engagement. Well-documented are various behaviours that disrespect boundaries between personal and professional lives, such as harassment, stalking, and other breaches of privacy. Yet, the rise of the ordinary, accessible star has also allowed greater exposure to previously hidden or stigmatised facets of figures’ lives, fostering safe spaces for perceived authenticity and vulnerability that can counter blind idealisation (6). Evolving industries & societies Under the diluted power networks of stardom today, we can describe celebrity image production as increasingly decentralised (6). Technology has made entry into the entertainment industry more accessible by providing numerous channels for artistic output, whether it be through releasing music independently on streaming services like Spotify, Bandcamp or Soundcloud, or creating short-form video skits on platforms like TikTok or Instagram. With top-down connections to age-old media institutions no longer required, the pool of faces that audiences can form relationships with has drastically expanded (7). Social norms – at the time of writing – have also welcomed the notion of diversified talents. As prevailing social, cultural and political structures shape value judgments made of stars (2), we have seen increased audience meaning-making in the dimensions of gender, ethnicity, class or sexual orientation over past decades (8) aligned with a gradual direction towards progressive and learned landscapes. Here, celebrity advocacy for causes and movements beyond the stage is nothing new, but fan bases can now dissect their forays into activism more publicly than ever before. A world unapologetically critical of “out of touch” (9) wealthy stars crooning out Lennon’s Imagine at the beginning of the pandemic would unlikely have welcomed the white-saviorist charity event that was Live Aid 1985 with as open arms as the dominant media narrative did then (10). A hyper-consumerist present If the exclusive stardom of yore can be likened to the dominance of a supermarket monopoly, then stardom today looks more like a diverse hub of online stores for buyers to ‘Click and Collect’ from. Whilst this setup offers diversified perspectives to a consuming audience, it embodies wider societal trends towards hyper-commodification. Market an image that sells well, and everyone will be famous for 15 minutes , as Andy Warhol supposedly declared (11). Reinforcing the ephemerality of mass consumerism are internet memes or trends (12) that morph and dilute rebellious celebrity motifs for overarching capitalistic agendas – think Brat Summer campaigns in the style of Charli xcx’s 2024 album co-opted by the most unethical multinational corporation you’ve ever come across. Like with the discourse exposing ‘nepo’ babies in the entertainment industry (13), we are reminded that despite the semblances of democratisation, the limelight remains far from a level stage. Stardom, beyond So what then? What lies in store for the future star? On one hand, the perception of proximity with the decline of ‘untouchable’ star personas can strengthen fan worship and deification, with frenzied consequences. On the other hand, increased artist-audience dialogue can pave the way for real change over performative gestures as lessening power imbalances bring a form of democratisation that can platform diverse and marginalised voices in art. All in all, stars today may no longer be able to fully present themselves and be perceived solely as spectral, enigmatic illusions that audiences can latch upon, but the new freedoms and avenues that come with being more truly known may be just as bedazzling. References 1. Robinson P. The great pop power shift: how online armies replaced fan clubs. The Guardian [Internet]. 2014 Aug 25; Available from: https://www.theguardian.com/music/2014/aug/25/great-pop-power-shift-how-online-armies-replaced-fan-clubs 2. Dyer R. Introduction. In: Heavenly Bodies [Internet]. Routledge; 2004. Available from: https://doi.org/10.4324/9780203605516 3. Wynarczyk N. Tracing Eastern Europe’s obsession with Depeche Mode [Internet]. Dazed. 2017. Available from: https://www.dazeddigital.com/music/article/36659/1/tracing-eastern-europe-s-obsession-with-depeche-mode 4. Hoffner CA, Bond BJ. Parasocial Relationships, Social Media, & Well-Being. Current Opinion in Psychology [Internet]. 2022 Feb;45(1):1–6. Available from: https://doi.org/10.1016/j.copsyc.2022.101306 5. Yodovich N. Buzzfeed’s “celebrities reading thirst tweets”: examining the sexualization of men and women in the #MeToo era. Journal of gender studies. 2024 Feb 28;33(8):1–11. Available from: https://doi.org/10.1080/09589236.2024.2324263 6. Driessens O. The Celebritization of Society and Culture: Understanding the Structural Dynamics of Celebrity Culture. International Journal of Cultural Studies [Internet]. 2013;16(6):641–57. Available from: https://doi.org/10.1177/1367877912459140 7. Carboni M. The digitization of music and the accessibility of the artist. Journal of Professional Communication [Internet]. 2014 Jun 4;3(2). Available from: https://doi.org/10.15173/jpc.v3i2.163 8. Stewart S, Giles D. Celebrity status and the attribution of value. European Journal of Cultural Studies [Internet]. 2019 Jul 21;23(1). Available from: https://doi.org/10.1177/1367549419861618 9. Caramanica J. This “Imagine” Cover Is No Heaven. The New York Times [Internet]. 2020 Mar 20; Available from: https://www.nytimes.com/2020/03/20/arts/music/coronavirus-gal-gadot-imagine.html 10. Grant J. Live Aid/8: perpetuating the superiority myth. Critical Arts [Internet]. 2015 May 4;29(3):310–26. Available from: https://doi.org/10.1080/02560046.2015.1059547 11. Nuwer R. Andy Warhol Probably Never Said His Celebrated “Fifteen Minutes of Fame” Line [Internet]. Smithsonian Magazine. Smithsonian Magazine; 2014. Available from: https://www.smithsonianmag.com/smart-news/andy-warhol-probably-never-said-his-celebrated-fame-line-180950456/ 12. Cirisano T. “Brat” summer and the dilemmas of going mainstream [Internet]. MIDiA Research. 2024. Available from: https://www.midiaresearch.com/blog/brat-summer-and-the-dilemmas-of-going-mainstream 13. Jones N. How a Nepo Baby Is Born [Internet]. Vulture. 2022. Available from: https://www.vulture.com/article/what-is-a-nepotism-baby.html Previous article Next article Enigma back to
- Death of the Scientific Hero
By Clarisse Sawyer < Back to Issue 3 Death of the Scientific Hero By Clarisse Sawyer 10 September 2022 Edited by Ruby Dempsey Illustrated by Quynh Anh Nguyen Next Trigger warning: This article mentions racism, sexism and misogyny and death. As a kid I was obsessed, like most kids, with animals of any kind. I would spend hours at a time scouring the beach for shells, getting sunburnt watching lizards, and tentatively feeding the praying mantises I caught, watching with morbid fascination as they hunted and dismembered the unfortunate crickets. It was only natural that I soon became interested in science. The long days of summer holidays were spent pouring over children’s encyclopaedias and watching David Attenborough documentaries. Through David Attenborough, I discovered two incredibly influential scientists - the co-discoverers of evolution, Charles Darwin, and Alfred Wallace. I idolised them, in particular, Wallace. As a shy child, who avoided the limelight like the plague, I had a natural inclination to root for the underdog, and Wallace was presented as such. Wallace was, in contrast to Darwin, much poorer, much more humble, and received much less credit for the theory of evolution than his co-discoverer Darwin. In my developing brain, Wallace took on the status of hero. I would chatter incessantly about him. I developed an interest in insects and butterfly collecting because he was a lepidopterist. I am sure my parents found me insufferable, but they hid their frustrations well, through subtle eye rolls and conversation changes, because they were happy to see me interested in science. So for my 11th birthday, my Dad bought me a book of Wallace’s letters from his time spent as a butterfly collector in the Malay Archipelago. The book was a lot drier than an 11 year old would have hoped for. Most of it was just taxonomy, peppered with the odd personalised comment complaining about the heat. But there was one passage which stood out to me in particular. A passage in which he describes shooting a “wild woman”, upon mistaking her for an orangutan in the forest canopy. In this section he details taking the baby she carefully carried on her back, and raising it as his own “n-word baby”. He promptly taxidermied the mother, with the intention of selling her remains to a wealthy private collector in England7. It was at this point I stopped reading. At 11, there was no way I could tell this was just an incredibly bad taste joke, and that in reality Wallace had actually shot a peculiar subspecies of orangutan, and not a Malaysian woman carrying her child. At 11, I believed my hero would kill me, if I wasn’t half white, if I wasn’t so light skinned, if I didn’t wear clothes, if I didn’t speak English. I would wonder for years afterwards: how brown would I have to be? To be plastinised, taxidermied, sold to some rich collector to sit in a sterile glass cabinet, at the back of some ex nobleman’s mansion. The passage ruined Wallace for me, but not science. Sometimes I wonder, if my passion for science was only marginally less, would I still be in science? I don’t know. For every child who is only mildly deterred by the racism or sexism of their former heroes, surely there is one child whose passion slowly fades, until the only time it is mentioned is by anxious mothers pushing their children to study medicine. I lost my hero, a precedent for who a scientist should be, in addition to developing a paranoia. A paranoia that if I were to start idolising another white, male, historical, scientific figure, I would be met with the same realisation that he would’ve despised me. And I haven’t been able to find a new hero since. Despite there being numerous people of colour, and women in science for a millennia before me, they weren’t the ones promoted to me, or if they were, I found them unrelatable save for their gender or the colour of their skin. They were people who were, 99% of the time, hard working to a fault, such as Marie Curie. Often this diligence was presented as being a detriment to their happiness. So my decision to study science, like many other women and people of colour, was also a decision to be my own precedent for what a scientist should be. While this is empowering, it is difficult not to envy those, like the privileged archetype of a white man, who might be able to draw confidence and inspiration from the figures in the preliminary pages of scientific textbooks. Whilst the majority of them may prove unrelatable, the sheer quantity would ensure that at least one would be a sympathetic character, in stark contrast to the singular, tokenistic entries on historical non-white or female scientists in such text books. But does it really have to be this way? Why should anyone have to feel alienated by scientific history? Why are there not more diverse heroes for us to fall back on? At the crux of my alienation from Wallace, and scientific history more generally, was deceit, more specifically what I perceived as lying by omission. The initial presentation of scientific figures such as Wallace by media, institutions and the like is so sympathetic and devoid of grisly details, that upon discovering the multifaceted nature of these individuals, I experienced a kind of historical whiplash. A scientific education is often presented as being objective. What you are taught in a classroom, at least at a primary or secondary level, is not meant to be subject to much nuance or interpretation. Now, when this concerns science itself, it is a non-issue, because it is true, for instance, that chromosomes are made of DNA, or that the first electron shell of an atom contains 2 electrons. The issue is that the perception of objectivity carries over into the way science history is taught. Unfortunately, this teaching is unavoidably subjective. Teachers and institutions often present positive anecdotes about scientists' hobbies and personal lives. A teacher may share for instance, an endearing fact about the influential French palaeontologist, Georges Cuvier, that he became as knowledgeable in biology as university trained naturalists by the age of 126. However, said teacher may neglect to mention the fact that after her death, Georges Cuvier dissected and taxidermied Sarah Baartman , a South African woman of the Khoisan tribe, and paraded her as a freak for the English public5. Her plastinated body remained on display at the Museum of Manin Paris until 19744. In this example, it would be impossible to say that the teacher’s presentation of Cuvier was objective. Choosing to share the nicest facts about a scientist, to make them appealing to your audience, while neglecting the ugly truths,is at best, irresponsible, and at worst, lying by omission. .Abhorrent actions, such as Cuvier’s treatment of Baartman’s corpse, a woman with whom he had danced and conversed with before her death, are treated as unnecessary details in objective scientific history, as they do not pertain to Cuvier’s scientific discoveries. However, equally unnecessary details, such as Cuvier’s early aptitude for biology, are peppered into school curricula liberally. However, it would be unfair to say that the primary reason why natural history is taught in this way is because of conscious racism and sexism. There are a multitude of explanations for why educators teach like this. Educators may choose to include only the nicer traits of scientific figures, in part perhaps because they do not want to risk disengaging students with affronting subject matter. Further, the morbidity and the racism of scientific history is not exactly appropriate content to teach to younger children. Precedent also plays a role in the way in which natural history is taught. Teaching natural history in an unbiased and inclusive fashion would require rewriting a lot of material. Educators would also have to reevaluate their own personal perceptions of historical figures, which is a difficult task. For instance in Australia, the textbooks A Short History of Australia2 and The Story of Australia3, which were staples of Australian high school history classes for decades, are white-centric stories of Australian exploration, which gloss over perturbing historic details such as massacres of Indigenous peoples. While teaching scientific history in a fair, unbiased and age appropriate manner might seem like an impossible task, there are a variety of small steps educators can take towards this end goal. A strong start would be the following; if teachers decide to include personal details about famous scientific figures, they should seek to include both positive and negative anecdotes, which frame negative actions in a disapproving light. The negative anecdotes serve to ensure that students don’t get ‘whiplash’ as they pursue their education, and also serve to show that modern science does not condone or approve of these actions. In the case of younger students, it is best for teachers to avoid talking about triggering topics, so teachers should teach scientific history from an objective standpoint sans personal details. Teachers also should, as part of their responsibilities as an educator, seek out alternative historical perspectives which challenge their own preconceived notions. And educational institutions should offer professional development courses which provide educators with a more balanced view on scientific history. These actions would help eliminate any subliminal biases teachers might have whilst teaching scientific history. And why are there not more diverse heroes for us to fall back upon? Lack of equal opportunity for marginalised groups in Western society for most of history and the systemic erasure of their contributions is an obvious reason, however through relying on secondary, colonial sources for information, instead of delving deeper into primary sources, educators and institutions inadvertently gloss over scientific contributions by marginalised groups. For example, the contributions of Indigenous Australian scientists and explorers are often ignored by museums. Many famous white explorers of Australia, such as Thomas Mitchell, Charles Sturt and Alexander Forrest worked closely alongside Indigenous guides, who helped navigate territory, and point out items of scientific interest, and their names are actually often acknowledged in primary sources1. For instance, one of explorer Thomas Mitchell’s chief guides, Yuranigh, is mentioned extensively in Mitchell’s personal accounts of his expeditions, and was acknowledged posthumously by Mitchell with a grave and monument1. These people, who were explorers in their own right, have largely been relegated to the footnotes of history and museums, in particular after the publications such as the aforementioned textbooks A Short History of Australia, and The Story of Australia in the 1950’s, which deliberately omitted Indigenous contributions to white Australian exploration in order to sell the false narrative of terra nullius. Luckily, through researching primary sources further, historians, educators and curators will be able to change the narrative, and shed light on these marginalised scientists. But what of scientific heroes? How is it possible to keep students engaged without the more personal aspects of science, given that many scientific figures will have to be cut from curriculums, at least for younger students?My answer to that would be to find new heroes. History is littered with people who made significant contributions without committing atrocities. And who knows, maybe in the void left by problematic figures, space could be cleared for more diverse heroes, the kind removed from history textbooks, such as Yuranigh; an exciting prospect. And yet, there is an unavoidable anguish in throwing out the old in favour of the new. Coming to terms with the fact that the people we idolised were terrible people is no easy feat. But all we can endeavour to do is to portray scientific figures as they were. To portray all aspects of these figures, good and bad, or none at all, and hopefully develop a new history, a new tradition, one that is inclusive, one for which everyone can be proud of and take solace in. References 1. Watson T. Recognising Australia's Indigenous explorers [Internet]. researchgate.net. 2022 [cited 19 May 2022]. Available from: https://www.researchgate.net/publication/321579451_Recognising_Australia's_indigenous_explorers 2. Scott E. Short History of Australia. Forgotten Books; 2019. 3. SHAW A. The story of Australia. London: Faber; 1975. 4. Parkinson J. The significance of Sarah Baartman [Internet]. BBC News. 2022 [cited 19 May 2022]. Available from: https://www.bbc.com/news/magazine-35240987 5. Kelsey-Sugg A, Fennell M. Sarah Baartman was taken from her home in South Africa and sold as a 'freak show'. This is how she returned [Internet]. Abc.net.au. 2022 [cited 19 May 2022]. Available from: https://www.abc.net.au/news/2021-11-17/stuff-the-british-stole-sarah-baartman-south-africa-london/100568276 6. Georges Cuvier [Internet]. Britannica Kids. 2022 [cited 19 May 2022]. Available from: https://kids.britannica.com/students/article/Georges-Cuvier/273885 7. Wallace A, Van Wyhe J, Rookmaaker K. Letters from the Malay Archipelago. Oxford: Oxford Univ. Press; 2013. Previous article Next article alien back to
- Why Our Concept of Colours is Broken | OmniSci Magazine
< Back to Issue 4 Why Our Concept of Colours is Broken by Selin Duran 1 July 2023 Edited by Tanya Kovacevic and Megane Boucherat Illustrated by Aizere Malibek The world that surrounds us is made from a combination of three main colours: red, yellow and blue. Known as the primary colours, it's the first thing we learn in primary school art class. In illusions, however, our concept of colours becomes warped and fails us. The only question is how do we fix it? Take the infamous colour-changing dress of 2015. This dress became an internet sensation due to its ambiguity of colour with the major question being “Is the dress black and blue or white and gold?” The dress, despite causing many online debates, is actually black and blue. Nevertheless this debate raises an important question about colours. Why do we see different colours in the same image? Let's begin with colour theory. Colour theory is a set of guidelines that artists use when mixing colours within the spectrum. With the intention of provoking different psychological responses, colours are used to either complement or contrast one another [1]. We see this through the infamous dress - with black and blue complimenting each, then gold and white. Our highly subjective perception allows us to see visually appealing combinations of colours juxtaposed to contrasting combinations. However, what we also need to consider are the light sources being used. Ranging from natural light to blue light and other artificial lighting, the light that we are exposed to can alter our perspective of colour. On our devices, we see colours through a series of red, green, and blue pixels that combine to make new colours for every image that we see [2]. Similarly, the frequent manipulation of our devices’ brightness also contributes to different colours being shown on the screens. These are the primary reasons why the famous dress was perceived so differently by everyone: each device shows a different version of the same colour depending on its display settings, which affects how many red, green and blue pixels there are. In addition to the colour theory, another effect— the Bezold Effect—is at its peak with the infamous dress. The Bezold Effect is an optical illusion where a colour’s appearance is affected by the presence of colours that surround the object [3]. For this dress, it’s seen through the shadows that form on and around the bodice. With brighter surroundings, such as the sun or an overly brightened screen, the blue from the dress appears gold to the eye, while the black appears white. The dress reverts to its original colours when the screen is darkened or artificial light is used. Circling back to colour theory, the changes in colours aren’t randomly allocated: they are opposing colours of the colour wheel. The wheel is a visual illustration of colours arranged by their wavelength, used to display the relationship of primary colours to their corresponding secondary colours [4]. With blue contrasting a yellow or gold, the changes in lighting perfectly display the contrasting colours on the wheel. The fascinating nature of colours is not something we can fix. In the era of digital displays and evolving technologies, we can’t see things the “right” way because there is no notable “right” or “wrong” way to look at the world. The dress is just one of those illusions that changes depending on the context and surroundings that it’s placed in. You can manipulate these colours and force them to change by physically changing the brightness on a device. So out of curiosity, I decided to conduct a little experiment of my own through an Instagram poll to see what my friends thought of this dress. While only 37 people participated, it was still fun to see what would happen with the votes; however, I was surprised to see the results after 24 hours. I expected a majority to choose the “real” colour of the dress, since the dress has been around in the media for a while and the answer is also online, but people still had contrasting opinions about the dress. With only 54% of people seeing black and blue and 46% white and gold, I began questioning our vastly different perceptions. The answer always seemed obvious as the dress was always black and blue not white and gold but that didn’t mean that other people saw what I saw. My favourite response came from a friend who saw the dress as blue and gold and after that, my opinion changed. For me, the dress is now blue and with tints of gold and I can’t see it any other way. This truly goes to show that there’s more behind the dress than what meets the eye. When I first saw the image my brightness was at the lowest it could possibly be and now after looking at the image enough, it’s just blue and gold. The ambiguity of this image is what makes the dress the best example of a real-life illusion. Other colour combinations act the same way in different lighting, but what we see is completely dependent on our perceptions, and every now and then, it’s always fun to put up a debate. References Eliassen MM. Colour theory. Salem Press Encyclopedia [Internet]. 2023 Jan 1 [cited 2023 May 13]; Available from: https://discovery.ebsco.com/linkprocessor/plink?id=30f4180b-d38d-38e6-95df-fcf469ab5c8a Mertes, A. (2021, February 23). Why Computer Monitors Display the Same Colors Differently . https://www.qualitylogoproducts.com/ . https://www.qualitylogoproducts.com/promo-university/why-monitors-display-different-colors.htm#:~:text=The%20pixels%20are%20in%20some,shows%20up%20on%20the%20screen Lasikadmin. (2022, June 2). What is Bezold Effect? | Useful Bezold Effect. LASIK of Nevada. https://lasikofnv.com/blog/test-your-vision-by-bezold-effect/#:~:text=What%20is%20the%20Bezold%20Effect,one%20to%20the%20human%20eye Understanding color theory: the color wheel and finding complementary colors . (n.d.). https://www.invisionapp.com/inside-design/understanding-color-theory-the-color-wheel-and-finding-complementary-colors/ Previous article Next article back to MIRAGE
- Message from the Editors in Chief
By Caitlin Kane, Rachel Ko, Patrick Grave, Yvette Marris Message from the Editors in Chief By Caitlin Kane, Rachel Ko, Patrick Grave, Yvette Marris 23 March 2022 Edited by the Committee Illustrated by Quynh Anh Nguyen Another year in science has passed, with 2022 disappearing into 2023. With a mandated return to campus life at the University, there seems a tangible break from the past three years of lockdowns, isolation and online existence. Over the summer holidays, four of our wonderful OmniSci contributers—Andrew, Julia, Lily and Yvette—have written about science that has made a mark in 2022, with topics spanning DNA of the ancient past to the future of art crafted by artificial intelligence. Our writers were supported by editors, Tanya and myself, and the cover and article art for this issue has been created by Quynh Anh. Thanks also goes to our behind-the-scenes events duo, Andrew (again!) and Aisyah, who have been working hard on promotion to showcase the work of our team on this mini-issue, and our treasurer-secretary, Maya, who keeps us all in line. On behalf of the whole team, we're incredibly excited to share our summer issue, 2022: A Year in Science. If you would like to support our work, you can sign up as a member, join our mailing list or get in touch at omniscimag@gmail.com—all this and more on our About Us page. Most importantly, please read on! Previous article Next article
- “Blink and you’ll miss it”: A Third Eyelid?
By Rachel Ko < Back to Issue 3 “Blink and you’ll miss it”: A Third Eyelid? By Rachel Ko 10 September 2022 Edited by Ashleigh Hallinan and Yvette Marris Rachel Ko Next The creature snarls a deep, thundering growl, tensing its protruding muscles that are covered in layers of thick, green, armour-like scales, individually rattling by the sheer force of its stance. Clenching its claws, the lizard glares with a bizarrely human expression, a villain trapped in a peculiar hybrid humanoid form. As the screams of terrified students fill the air, the camera zooms into the mutant’s glistening yellow eye, and it blinks; a slimy, translucent covering flickers across its eyeball, leaving a trail of moisture - grotesque proof of its reptilian form. A charm of the cinematic world is that aliens, radioactive spider superheroes and giant mutant lizards can exist in the same universe as the regular person. On a recent movie night, watching The Amazing Spiderman, the villain Lizard caught my eye. The creature is a metamorphosed version of human scientist Dr Curt Connors, who had attempted cross-species genetic regeneration on himself. Largely CGI, the Lizard’s primitive no-frills characterisation makes him an unconventional superhero antagonist. However, upon focus, these exaggerated reptilian characteristics are wha become staples of the Lizard’s uniquely villainous appeal: the alien-green colouring, the razor-sharp claws, the terrifying teeth and, of course, the glistening yellow eyes. Figure 1: Spiderman's 'The Lizard' In reference to the creation of these eerie eyeballs, animation supervisor David Schaub confirmed the purposeful inclusion of a nictitating membrane (1). This membrane is a slimy skin-like covering more commonly known as the Third Eyelid. In animals such as birds, reptiles, fish, amphibians, and some mammals (2), it acts as a bizarre protective mechanism that maintains moisture while retaining vision (3) - and also gives the Lizard’s glare that extra kick. Acting like a windscreen wiper, the membrane ‘nictitates’, meaning it blinks, to keep debris and dust out of the eye while simultaneously hydrating it. Its transparency also allows vision underground or underwater (4). Figure 2: A bird blinking! There is just one primate species known to have a prominent nictitating membrane: the Calabar angwantibo, also known as the golden potto, which is a rare African prosimian primate found only in Cameroon and Nigeria (5). Figure 3: Look at the Calabar's nictating membrane! The membrane is a major characterising feature of The Amazing Spiderman’s creepy mutant reptilian aura. However, this Third Eyelid actually has a homologous counterpart in Dr Connors’ eyes too. In fact, it is found in all humans, and is known as our plica semilunaris, a vertical fold of conjunctiva in the inner corner of the eye (6). Although it plays a minor role in eye movement and tear drainage (7), the plica semilunaris has nowhere near as great a function in humans as the nictitating membrane does in animals (8). The plica semilunaris and its associated muscles are merely an evolutionary remnant of the nictitating membrane that existed in our reptilian ancestors millions of years ago (9). Evolution is driven by selective advantage: the traits that allow organisms to survive and reproduce are the ones that are selected for and thrive within the population, passed down from one generation to the next (10). Traits that are disadvantageous to organisms decrease their chance of survival and reproduction, meaning fewer offspring will inherit the trait, causing it to eventually disappear from the population (11). The mystery remains as to why human ancestors lost the nictitating membrane in the first place, but it is likely that changes in habitat and lifestyle regarding eye physiology made it selectively advantageous to lose the Third Eyelid, rather than wasting precious energy on maintaining a no-longer-vital mechanism (12). For some reason, though, once the nictitating membrane had evolved into nothing more than a miniscule pink fold in the corner of the eye, it still persisted. Some argue that this is because humans have had no evolutionary incentive to completely lose them (13) – the plica semilunaris is just harmless enough that it has flown under the radar of evolution’s cut. Having suggested that, however, the primary clinical significance of the plica semilunaris has been connected to allergies of the eye, in which release of inflammatory molecules like histamine causes the tissue to become swollen and itchy (14). Thus, it is worth considering another argument: that the persistence of the plica semilunaris may be indicative of some beneficial function, particularly in its role in human eye protection. It has been found that the tissue observed in early intrauterine (within the uterus) development has a dense infiltration of immune cells like macrophages and granulocytes that serve to engulf and destroy foreign invaders of the tissue (15). Along with the abundance of blood vessels and immune chemical signalling, this has suggested a specialised role in eye protection, a benefit that may have very well ensured the plica semilunaris’ survival within human populations until this day (16). One fascinating clinical case, which showcases the outlandish capabilities of this vestigial feature, is of a child for whom it was not a question of why the plica semilunaris persisted, but an actual nictitating membrane. This peculiar instance was presented on a 9 year-old girl whose left eye had a non-progressive translucent membrane covering it horizontally. The globe of the eye was able to move freely beneath the membrane, suggesting that there was no attachment. However, it was causing amblyopia (also known as a lazy eye), and poor vision, so the nictitating membrane was successfully removed by simple excision (17). Figure 4: The plica semilunaris Figure 5: A clinical case of a human nictating membrane The only other recorded case of persisting nictitating membrane was an infant boy born prematurely with Edwards syndrome, who had nictitating membranes in both eyes (18). However, due to the baby’s infancy and condition, membrane imaging was unobtainable. Thus, arguably, the most striking aspect of the 9 year-old girl’s case was the pre-procedure imaging of her eye: an intriguing, almost alien-like fusion of the human eye and that of our reptilian ancestors. This case study can be interpreted as an exaggerated example of an existing link between the nictitating membranes we see in animals today, and the plica semilunaris that exists, tucked away, in the corner of our very own eyes. So, next time you find yourself staring into your partner’s baby blues, or putting on eyeliner in the mirror, keep an eye out for this fascinating evolutionary remnant; but be quick because - blink and you’ll miss it. References Sarto D. 'Spider-Man'’s Lizard Part 1: The Animation [Internet]. Animation World Network. 2012 [cited 4 May 2022]. Available from: https://www.awn.com/vfxworld/spider-mans-lizard-part-1-animation Butler A, Hodos W. Comparative vertebrate neuroanatomy. Hoboken (New Jersey): Wiley-Interscience; 2005. Why do cats have an inner eyelid as well as outer ones? [Internet]. Scientific American. 2006 [cited 4 May 2022]. Available from: https://www.scientificamerican.com/article/why-do-cats-have-an-inner/ The Equine Manual [Internet]. Elsevier; 2006. Available from: http://dx.doi.org/10.1016/B978-0-7020-2769-7.X5001-1 Montagna W, Machida H, Perkins EM. The skin of primates. XXXIII. The skin of the angwantibo (Arctocebus calabarensis) [Internet]. Vol. 25, American Journal of Physical Anthropology. Wiley; 1966. p. 277–90. Available from: http://dx.doi.org/10.1002/ajpa.1330250307 Plica semilunaris [Internet]. Merriam-Webster.com medical dictionary. [cited 4 May 2022]. Available from: https://www.merriam-webster.com/medical/plica%20semilunaris LaFee S. Body and Whole [Internet]. UC Health - UC San Diego. 2016 [cited 4 May 2022]. Available from: https://health.ucsd.edu/news/features/pages/2016-06-30-listicle-body-and-whole.aspx Dartt D. Foundation Volume2, Chapter 2. The Conjunctiva–Structure and Function [Internet]. Oculist.net. 2006 [cited 4 May 2022]. Available from: http://www.oculist.net/downaton502/prof/ebook/duanes/pages/v8/v8c002.html Gonzalez R. 10 Vestigial Traits You Didn't Know You Had [Internet]. Gizmodo. 2011 [cited 4 May 2022]. Available from: https://gizmodo.com/10-vestigial-traits-you-didnt-know-you-had-5829687 Sukhodolets V. V. (1986). K voprosu o roli estestvennogo otbora v évoliutsii [The role of natural selection in evolution]. Genetika, 22(2), 181–193. Sukhodolets V. V. (1986). K voprosu o roli estestvennogo otbora v évoliutsii [The role of natural selection in evolution]. Genetika, 22(2), 181–193. Gonzalez R. 10 Vestigial Traits You Didn't Know You Had [Internet]. Gizmodo. 2011 [cited 4 May 2022]. Available from: https://gizmodo.com/10-vestigial-traits-you-didnt-know-you-had-5829687 Kotecki P, Olito F. We No Longer Need These 9 Body Parts [Internet]. ScienceAlert. 2019 [cited 4 May 2022]. Available from: https://www.sciencealert.com/we-no-longer-need-these-9-body-parts Bielory L, Friedlaender MH. Allergic Conjunctivitis [Internet]. Vol. 28, Immunology and Allergy Clinics of North America. Elsevier BV; 2008. p. 43–58. Available from: http://dx.doi.org/10.1016/j.iac.2007.12.005 Arends G, Schramm U. The structure of the human semilunar plica at different stages of its development a morphological and morphometric study [Internet]. Vol. 186, Annals of Anatomy - Anatomischer Anzeiger. Elsevier BV; 2004. p. 195–207. Available from: http://dx.doi.org/10.1016/S0940-9602(04)80002-5 Arends G, Schramm U. The structure of the human semilunar plica at different stages of its development a morphological and morphometric study [Internet]. Vol. 186, Annals of Anatomy - Anatomischer Anzeiger. Elsevier BV; 2004. p. 195–207. Available from: http://dx.doi.org/10.1016/S0940-9602(04)80002-5 Vokuda H, Heralgi M, Thallangady A, Venkatachalam K. Persistent unilateral nictitating membrane in a 9-year-old girl: A rare case report [Internet]. Vol. 65, Indian Journal of Ophthalmology. Medknow; 2017. p. 253. Available from: http://dx.doi.org/10.4103/ijo.IJO_436_15 García-Castro JM, Carlota Reyes de Torres L. Nictitating Membrane in Trisomy 18 Syndrome [Internet]. Vol. 80, American Journal of Ophthalmology. Elsevier BV; 1975. p. 550–1. Available from: http://dx.doi.org/10.1016/0002-9394(75)90228-7 Images Figure 1: Galloway, R. (2022, January 25). Lizard originally had a different look in 'Spider-Man: No way home'. We Got This Covered. Retrieved August 9, 2022, from https://wegotthiscovered.com/movies/lizard-originally-had-a-different-look-in-spider-man-no-way-home/ Figure 2: Hudson T. (2010, July) Retrieved Sep 13, 2022, from https://en.wikipedia.org/wiki/Nictitating_membrane#/media/File:Bir d_blink-edit.jpg Figure 3: Sharma R. Calabar angwantibo - Alchetron, The Free Social Encyclopedia [Internet]. Alchetron.com. 2018 [cited 7 May 2022]. Available from: https://alchetron.com/Calabar-angwantibo Figure 4: Amir, D. (2019, January 16). Twitter. Retrieved August 9, 2022, from https://twitter.com/dorsaamir/status/1085557444196 081664 Previous article Next article alien back to
- A few words on (Dis)Order! | OmniSci Magazine
< Back to Issue 2 A few words on (Dis)Order! From modelling the spread of COVID-19 to analysing gene sequences, science has its way of providing clarity and order in situations of apparent chaos. Our Editors-in-Chief give their take on Issue 2’s theme of (Dis)Order, in their various fields of study. by Sophia, Maya, Patrick and Felicity 10 December 2021 Edited by the Committee Illustrated by Jess Nguyen Rainbow cars, erratic robots, and a circuit named Chua — Sophia Lin In Grade 10, I pressed ‘Play’ on my computer, and was captivated by the turbulent air flowing around my race car, rendering the screen with a rainbow of colours. This was the first time I had encountered a tool called Computational Fluid Dynamics, commonly used to analyse the aerodynamics of systems. Turbulent air is probably the most textbook example of chaos, their motion described by the notorious Navier-Stokes equations. But chaotic systems exist everywhere in the natural world and accounting for them in models is essential to be able to test and improve our engineering designs. But how can we use chaos? In 2001, researchers Akinori Sekiguchi and Yoshihiko Nakamura first suggested applying chaotic systems to path planning of robots. [1] Later on, researchers Christos Volos et al. applied the Arnold chaotic system to two active wheels of a simulated mobile robot, allowing it to completely, and quickly, scan the unknown terrain in an erratic, unpredictable way. [2] This exploration strategy is not new in nature, however, with research suggesting that ants partly use random motion to search areas for food. [3] Finally, can we engineer chaos? In the field of electrical engineering, it turns out that this is pretty simple! Chua’s circuit contains your standard electrical components - just a linear resistor, two capacitors, one inductor, and a special non-linear resistor called “Chua’s diode” [4] , and is able to generate a funky “double-scroll” pattern which never repeats. The applications are just as exotic, ranging from communication systems, brain dynamics simulations and even music composition! It’s apparent that learning to model, imitate and harness chaos is key to engineering for our (dis)orderly world. Computer simulation of Chua’s circuit [5] Chua’s Circuit diagram [5] The Chaos in Communication — Maya Salinger Throughout the animal kingdom, and particularly amongst humans, communication methods are continually evolving for structures to be as efficient as possible. [6] In relation to human languages, there are of course thousands of languages being spoken worldwide everyday. It would not surprise me if you said that it was a daily occurrence for you to hear a conversation in a language you could not even remotely understand. To your untrained ears, these languages’ sounds, vocabulary and intonation patterns would be unfamiliar, with the combination of these structures sounding very chaotic. However, languages are inherently very structured due to their natural inclination towards efficiency. This structure is observed in hundreds of ways, from the patterning of the tiniest units of sounds, known as phonology, to the much larger structure of phrases and sentences, known as syntax. However, each language has its own unique set of structures, thus explaining their diversity and our inability to comprehend unfamiliar languages. Furthermore, structure in communication is not limited to human language. Throughout the animal kingdom, there are many species that consciously order certain movements or sounds to express particular information. For example, honeybees have a refined method of communication called a “waggle dance”. [7] Whilst it appears to you or I that a honeybee’s movements are random, they strategically encode the precise distance and direction of a nearby flower patch. Structured communication can be seen widely throughout the animal kingdom, despite how chaotic it can appear on the surface for those outside the language community. Our Bodies, in Chaos — Felicity Hu Like it or not, we are no strangers to disorder. In the changing world around us, chaos seems to be wherever we look: from our unpredictable Melbourne weather to the many phases of disarray brought on by COVID-19. Although we might encounter disorder in our external environment, we also carry around a little chaos of our own, packaged unassumingly within our bodies. What better example than in our own heads? Our brains have an astonishing number of around 86 billion neurons [8], polarising and depolarizing at different rates [9] The chaos of our neural network, with its many components phasing in and out of firing, its cells cycling through life and death, happens even as you are reading this. From the chaos of our brains, however, comes the clarity and processes we use every day. When preparing a cup of tea for a study break, for example, the chaos in our brains follows the wandering of our minds as we wait for the water to boil. Even after we have a steaming cuppa on our table, our ability to learn the wild and wonderful things from our university textbooks arises from the tangle of neurons and signals in our brains. While we aim to control the chaos in the world around us, sometimes it is worth appreciating the fact that we, too, have chaos in our own minds. And even more astoundingly, that we can derive clarity from it. Learning to Count — Patrick Grave I was never very good at counting. As a tiny boy I sat cross-legged, thumbing through the strands of my frayed shoelace, when I finally figured out how to count by twos. Until this point in Grade One, I did not know how I did addition; maybe I copied from the kid next to me, or perhaps there was something greater. See, on the list of important human inventions, counting ranks fairly highly. It takes a mysterious instinct, that of ‘more’ and ‘less’,and formalises it, creating order and power. When ancient peoples began using clay tokens with numeric values [10] and writing symbols on tablets [11], they could move beyond the four objects kept in visual memory [11] or the ten kept on fingers. They could track larger quantities: people, livestock, and wealth. [12] [15]: Ancient Uruk accountancy tokens and protective seal [16]: Counting using tally marks on sign at Hanakapiai Beach As a 10-year-old, I would tally things on my legs with Sharpie: Tennis serves, laps of the oval, footy goals for the season. Mum was not impressed. Over time, numbers branched out. Arithmetic was invented. Greek scholars like Archimedes used negative powers to store fractional parts [13]. In the Hindu-Arabic system, the number zero exists, and each digit’s position matters, allowing for efficient computation. This paved the way for banking, finance, and modern industry [14]. My friend showed me fractions a year early. With hushed tones and nervous side-glances, he wrote one number over another. They still feel a bit like magic. While modern maths has largely preserved the Hindu-Arabic system, other ways of counting have existed, each tailored to a civilisation’s needs. The Incas kept numerical records using knots in rope as they were less interested in advanced computation [15]. The Maya peoples used a base-20 system. [16] So, these numbers and counting systems are not natural. Instead, they have been imposed on nature by the machine of human progress. Counting tells a rich story of human development and of each civilisation’s place in that rich tapestry. Unlike humanity, I’m still not very good at counting. To our team and our readers We’d like to extend a massive thank you to the team behind Issue 2 of OmniSci Magazine! It has been a hectic, but rewarding few months, and we are so grateful for the effort, care and passion that has brought this issue together. We can’t wait to reflect on our journey so far, and bring more science to our readers in 2022. References Nakamura, Yoshihiko, and Akinori Sekiguchi. “The Chaotic Mobile Robot.” IEEE Transactions on Robotics and Automation 17, no.6 (Dec 2001): 1-3. http://projectsweb.cs.washington.edu/research/projects/multimedia5/JiaWu/review/Cite1.pdf Volos, Christos, Nikolaos Doukas, Ioannis Kyprianidis, Ioannis Stouboulos and Theodoros Kostis, Chaotic Autonomous Mobile Robot for Military Missions (Rhodes Island, Proceedings of the 17th International Conference on Communications, 2013), 1-6, Garnier, Simon, Maud Combe, Christian Jost, Guy Theraulaz. “Do Ants Need to Estimate the Geometrical Properties of Trail Bifurcations to Find an Efficient Route? A Swarm Robotics Test Bed.” PLoS Computational Biology 9, no.3 (2013): doi: 10.1371/journal.pcbi.1002903 Gauruv Gandhi, Bharathwaj Muthuswamy, and Tamas Roska, “Chua’s Circuit for High School Students”, Nonlinear Electronics Laboratory, https://inst.eecs.berkeley.edu/~ee129/sp10/handouts/ChuasCircuitForHighSchoolStudents-PREPRINT.pdf Shiyu Ji, “ChuaAttractor3D”, published November, 2016, https://en.wikipedia.org/wiki/Chua%27s_circuit#/media/File:ChuaAttractor3D.svg Gibson, Edward, Richard Futrell, Steven T. Piandadosi, Isabelle Dautriche, Kyle Mahowald, Leon Bergen, Roger Levy, “How Efficiency Shapes Human Language,” CellPress 23, 5 (2019): 389-407, https://doi.org/10.1016/j.tics.2019.02.003 . Landgraf, Tim, Raúl Rojas, Hai Nguyen, Fabian Kriegel, Katja Stettin, “Analysis of the Waggle Dance Motion of Honeybees for the Design of a Biomimetic Honeybee Robot,” PLoS ONE 6, 8 (2011): e21354, https://doi.org/10.1371/journal.pone.0021354 . Azevedo, Frederico A.C., Ludmila R.B. Carvalho, Lea T. Grinberg, José Marcelo Farfel, Renata E.L. Ferretti, Renata E.P. Leite, Wilson Jacob Filho, Roberto Lent, and Suzana Herculano-Houzel. 2009. "Equal Numbers Of Neuronal And Nonneuronal Cells Make The Human Brain An Isometrically Scaled-Up Primate Brain". The Journal Of Comparative Neurology 513 (5): 532-541. doi:10.1002/cne.21974. Kalat, James. 2018. Biological Psychology. Mason, OH: Cengage. Schmandt-Besserat, Denise. 2008. "Two Precursors Of Writing: Plain And Complex Tokens - Escola Finaly". En.Finaly.Org. http://en.finaly.org/index.php/Two_precursors_of_writing:_plain_and_complex_tokens . Schmandt-Besserat, Denise. 1996. How Writing Came About. Austin: University of Texas Press. Finn, Emily. 2011. "When Four Is Not Four, But Rather Two Plus Two". MIT News | Massachusetts Institute Of Technology. https://news.mit.edu/2011/miller-memory-0623 . Law, Steven. 2012. "A Brief History Of Numbers And Counting, Part 1: Mathematics Advanced With Civilization". Deseret News. https://www.deseret.com/2012/8/5/20505112/a-brief-history-of-numbers-and-counting-part-1-mathematics-advanced-with-civilization . Archimedes, and Thomas Heath. 2002. The Works Of Archimedes. New York: Dover. "The Use Of Hindu-Arabic Numerals Aids Mathematicians And Stimulates Commerce | Encyclopedia.Com". 2021. Encyclopedia.Com. Accessed December 9. https://www.encyclopedia.com/science/encyclopedias-almanacs-transcripts-and-maps/use-hindu-arabic-numerals-aids-mathematicians-and-stimulates-commerce . Bidwell, James K. 1967. "Mayan Arithmetic". The Mathematics Teacher 60 (7): 762-768. doi:10.5951/mt.60.7.0762. Nguyen, Marie-Lan. 2009. Accountancy Clay Envelope Louvre Sb1932.Jpg. Image. https://commons.wikimedia.org/wiki/File:Accountancy_clay_envelope_Louvre_Sb1932.jpg . War, God of. 2010. Hanakapiai Beach Warning Sign Only. Image. https://commons.wikimedia.org/wiki/File:Hanakapiai_Beach_Warning_Sign_Only.jpg . Previous article back to DISORDER Next article
- A Message from the Editors in Chief | OmniSci Magazine
< Back to Issue 5 A Message from the Editors in Chief Rachel Ko & Ingrid Sefton 24 October 2023 Edited by Committee Illustrated by Aisyah Mohammad Sulhanuddin “There are known knowns. These are things we know that we know. There are known unknowns. That is to say, there are things that we know we don't know. But there are also unknown unknowns. There are things we don't know we don't know.” - Donald Rumsfeld Science should never be considered as pursuing absolute truth. In fact, more often than not, the deeper we dive into its exploration, the more questions that arise. The world of science affords us choices in how we appropriate the understandings and knowledge gained in its study. Every day, science pushes us to tiptoe this fine line between pushing boundaries and crossing them altogether. It is perhaps this unknown that makes the pursuit of science so wicked in itself, taunting us with the promise of making the next big discovery, or finally finding the cure to cancer. But it is also what drives us, entrances us, and keeps our desire for knowledge burning — it’s edge-of-your-seat exciting. At its onset, we envisioned this issue as a chance to probe the mysterious nuances of science — a peek into the ‘Wicked’ness of the world. Seeking to ask questions of the ethical, the malicious and the unknown, contributors were inspired to delve into the darker sides of science. Each article ventures into the limits of what we do, and, just as importantly, don’t know in this ever-evolving field. The word Wicked in itself is a complex character, begging for ambiguous interpretation. Is there such a thing as pure evil? Are we all, just a bit, inherently wicked? What makes something wickedly cool? (Was Kristin Chenoweth’s Glinda the best portrayal that Broadway could ask for?) And so, in the hands of our creators, something wicked this way comes … As with every edition of our magazine, each piece has been created, edited and illustrated entirely by students. This issue continues to stand true to our aim of providing a platform within, and beyond, the university community for students of all backgrounds to craft their science communication skills in a supportive, creative environment. Countless hours have been poured into the curation of each edition with the hope of making innovative science content easily accessible — so please, enjoy! To all our passionate, dedicated contributors - thank you for the time you have invested in crafting the wonderful, wicked world of Issue 5 of OmniSci. It has been a privilege to watch the collaboration of inquisitive minds, from diverse scientific and artistic worlds, produce this collection of work. We also wish to extend our gratitude to you, our wonderful readers, in your ongoing support of OmniSci. The time you give to reading and engaging with our student-driven magazine does not go unnoticed, motivating and inspiring us for our future endeavours. Now, take a moment, and come venture into the Wicked world of Issue 5 with us… Wicked back to
- Hidden Worlds: a peek into the nanoscale using helium ion microscopy | OmniSci Magazine
< Back to Issue 2 Hidden Worlds: a peek into the nanoscale using helium ion microscopy How do scientists know what happens at scales smaller than you can see using an optical microscope? One exciting method is the helium ion microscope which can be used to view cells, crystals and specially engineered materials with extreme detail, revealing the beauty that exists at scales too small to imagine! by Erin Grant 10 December 2021 Edited by Jessica Nguy and Hamish Payne Illustrated by Erin Grant The room is white, with three smooth walls and a fourth containing a small sample prep bench and high shelves. In the centre is a desk with three monitors. Next to it, occupying most of the space, is the microscope. Eight feet tall, a few feet wide, resting on an isolated floor surrounded by caution tape; “NO STEP” written in big block letters. Wires protrude from its tiered shape in orderly chaos. It is a clean, technological space; we are ready to explore science. A colleague and I are at the Materials Characterisation and Fabrication Platform of the University of Melbourne to finish off the last steps of a scientific paper I’ve been working on for many years. What I need, as the icing on the cake, is an image. What does my sample look like way down there, at the nanometre scale? Objects that are only nanometres in size are very hard to imagine when we’re used to thinking about metres, centimetres, or maybe even millimetres. We can see those length scales; they are part of our everyday. So, if you’re told that proteins have a diameter of a few nanometres, what does that mean? Well, to be precise, a nanometre is one-billionth of a metre. A human hair, the go-to yardstick for describing small things, has a width between 0.05-0.1 millimetres, which means that if you wanted to slice a hair into nanometre-wide strands you’d end up with nearly 100,000 pieces. Unfortunately, that’s still hard to visualise, but I’ve found that when working with and thinking about scales like this every day, you gain a sort of mental landscape that small things occupy, perhaps not entirely in context, but a space that contains an overall ‘vibe’ of smallness. I first noticed this when I worked in a laboratory that studies the tiny nematode worm C. elegans. These creatures are half a millimetre long, so although they are clearly visible to the naked eye, you need a microscope if you want to use them for science. After looking at these tiny creatures under magnification for many weeks, I came to recognise a feeling almost like being underwater. Upon putting my eyes to the lens, my focus would change from the macroscopic world around me, to one of minutiae. This change in perspective was quite immersive, I almost felt like I was inhabiting that small petri dish too. Working with samples even smaller than that now, I have carried some of that mental landscape with me. It now feels commonplace to imagine tiny systems, such as crystals or molecules which were once foreign. Much of this ability to visualise small things comes from the fact that in many cases, we can actually see them too. Physics has given us many tools with which we can peer into the smallest systems that exist. Helium ion microscopy, which I have come here to carry out, is one such technique. Dr Anders Barlow runs the helium ion microscope (HIM) at this facility. He warmly welcomes me and my colleague into the quiet room and jumps straight into an enthusiastic explanation of the machine – he can tell we’re not just here for some pictures, we want to know the inner workings of the microscope too. The HIM is a bit like the more mature surveyor of minuscule worlds: the electron microscope. While a regular optical microscope uses light to illuminate a sample, the electron microscope uses electrons. When they collide with the sample these electrons can bounce off or lose energy through several mechanisms. The lost energy can go into heat or light, but more usefully, the energy might be transferred to other electrons in the sample, called secondary electrons, ejecting them like a drill removing rocks from a quarry. The secondary electrons can be detected at each point across the sample as the beam is scanned over its surface. If more electrons are detected, then the pixel at that point is brighter compared to areas where there are fewer electrons. This tells you about the topography or composition of the sample at that point on its surface and provides a grayscale image. The HIM works in the same way, but it can generate sharper images because helium ions are heavier than electrons. This is important because the increased resolution of electron and helium ion microscopes is enabled by their quantum mechanical properties - namely the particle’s wavelength. You may have heard about the wave-like nature of light, which is a basic property of quantum mechanics. Particles also have a wavelength, called the de Broglie wavelength, which is inversely proportional to their mass - the heavier the particle, the shorter the wavelength. Having a shorter wavelength allows smaller details to be resolved because of a pesky phenomenon called diffraction. Diffraction occurs when a wave encounters a gap that is of the same or smaller width to its wavelength. When this happens, the wave that emerges on the other side will be spread out. You can think of the features that you want to image as being similar to gaps, so when light, or a particle, interacts with features that are very close together it will spread out, making those features blurry or even invisible. But if you can ensure that the wavelength is smaller than whatever feature you want to see, diffraction will not occur. Interestingly, physicists can actually take advantage of diffraction, and another phenomenon called interference, when they study periodic structures like crystals, but that’s a different article! So, because the de Broglie wavelength is very short for particles with mass, like electrons, an electron microscope can generate images of higher resolution than an optical microscope. Likewise, helium ions are even heavier than electrons because they are composed of one electron, two protons, and two neutrons. This makes them about 7,000 times heavier than a single electron (electrons are very light compared to protons and neutrons!) and consequently the images they can make are very sharp. With our samples ready, lab manager Anders loads my sample into the microscope and begins lowering the pressure in its internal chamber. Having a high vacuum – approximately a billion times lower than atmospheric pressure – is essential because it prevents air from interfering with the helium beam. Making the beam is perhaps the most miraculous part of this technological feat. At the very top of the microscope’s column, there’s a tiny filament shaped like a needle. Not like a needle, in fact, it is the sharpest needle we humans can make. To achieve this, the point is shaped by first extreme heat, and then some extreme voltages until the very tip is composed of only three atoms, reverently referred to as the trimer. Once the trimer has been formed, a high voltage is applied to the needle, resulting in an extreme electric field around the tip. Next, helium gas is introduced into the chamber and individual helium atoms are attracted towards the region of the high electric field. The field is so strong that it strips each helium atom of one electron, ionising it, and these now positively charged ions are repelled from each of the three atoms in the trimer as three corresponding beams. Using sophisticated focusing fields down the length of the column allows Anders to choose only one of the beams for imaging; we are creating a picture using a beam only one atom wide! Generating such a precise beam requires constant maintenance, but once Anders is satisfied with how it looks today, he begins scanning over a large area for what we’ve come to find: tiny proteins stuck to a diamond. In an experimental PhD, you often find yourself answering small incremental questions and today I want to know how well I’ve attached these proteins to my diamond and what the coverage looks like. Other measures have told me that I probably have a lot of them, but the best way to know is to have a look! That’s what Anders does for researchers at the university; he helps us find out whether we have done a good job putting things together or coming up with new techniques. This is something he loves about his job. “I love the exposure I get to many areas of science,” he says, “Imaging of all forms is ubiquitous in research, and the HIM is applicable to most fields, so we see samples from materials science, polymers, nanomaterials, and biomaterials, through to medical technologies and devices, to cell and tissue biology of human, plant and animal origin. I never get tired of seeing what new specimens may come through the lab door.” Unfortunately, the first images we see are very dark and washed out, like a photograph taken in low-light; not many secondary electrons are making it to the detector. To combat this, Anders uses a flood gun to stop charge build up on the surface of the diamond. When the helium ions create secondary electrons, they are ejected from the surface at low speeds. As electrons are negatively charged, the bombarded surface, which now lacks electrons, will become positive and the low energy secondary electrons will be attracted back to the surface instead of making it to the detector. In an electron microscope this is avoided by coating insulators, such as my diamond, with a conductive material like gold. If the surface is conductive, the positive charge that is left behind by the secondary electrons will be offset by electrons from the metallic coating that can flow towards the sudden appearance of positive charges. In this case, the ejected electrons can escape and be detected. However, a coating like this would reduce the resolution of the image; if you want to measure proteins that are twelve nanometres high, but you put a three-nanometre coating over them, you’ll lose a lot of the resolution! To get around this, the HIM uses the flood gun, which lightly sprays the surface with electrons of low energy as the helium beam passes over. This neutralises the surface and lets the secondary electrons escape in the same way as having a conductive layer. Once Anders turns on the flood gun, the contrast increases, allowing us to zoom in on a small region of the diamond, and there they are! Thousands of spherical proteins arranged neatly across the surface, only twelve nanometres in diameter. The sight is spectacular, only one try and we got what we came for. I am three years into a PhD and I’ve become very used to the feeling of disappointment that can accompany new experimental techniques. Things rarely work out the first time around, so to see those little spheres straight away was magical. Dotted across the diamond surface is another, extra, gem. To keep protein nice and happy, you must prepare it in a salty solution. So, when the protein was deposited, some regular table salt, NaCl, came too. We can see this salt in our images as crystals in two distinctive and very beautiful patterns which you can see in the images below. Protein on the surface of my diamond. Each small pale circle is one of these spherical proteins. The first image shows a large creeping pattern, reminiscent of snowflakes or tree roots, which spreads its soft fingers across several hundred nanometres. These crystals have taken on an amorphous pattern, where the crystal structure is broken up rather than being one continuous arrangement of the atoms. The second pattern however, shown in the right image, is what a continuous NaCl crystal looks like. When large enough crystals can form without becoming amorphous they look like precise cubes of various sizes all strewn about. One of my favourite aspects about looking at very small things, is how the patterns you see often mirror those at much larger scales. Look at a fingerprint and you’ll find mountains and valleys, or the roots of a tree and you’ll see a river system. Salt (NaCl) can take on a highly ordered structure shown by the cubic crystals (left) or an amorphous pattern similar in shape to tree roots (right). The astonishing images we get from this single session are all in a day’s work for Anders. He has imaged numerous kinds of cells on all manner of interesting substrates, patterned surfaces covered in needle-like protrusions, and many kinds of man-made materials. Today, there are vials on his prep-bench which, at first glance, look much like jars of hair. However, they are not hair, in fact they are strands of carbon fibre covered in various coatings, awaiting examination. ‘What are your favourite types of samples to look at?’ I want to know. “Cell biology is fascinating,” he says. “We’ve imaged red blood cells, pancreatic cells, stem cells, and various bacterial cells in this microscope. Most often researchers are interested in cell life and death, and the HIM assists by providing high resolution images of the structure and surface topography of the cell membrane.” Recently however, Anders has been helping researchers look at polymer materials for water filtration. “These are hierarchical porous structures, meaning they’re engineered to have pore sizes that vary through the membrane. It is stunning to see the materials at low magnification with large pores, and as we zoom in and in and in, to see new pore sizes become visible at each level, like a material engineered with a fractal quality.” One of the unique things about the HIM, Anders reminds me, is that it’s not just for imaging. Since helium ions are heavy, they carry a higher momentum than electrons. “We leverage the momentum of the ions to actually modify structures too. We can create new surface properties, new devices, new technologies, on a scale that is often too small for any other fabrication technique. This is some of the most exciting work.” If you know anyone who needs some nanoscale drilling done, then the HIM is your instrument! Today’s excursion across the university campus has been thrilling. I got what I came for and I’m excited to find other projects that could benefit from the insight and beautiful images the HIM can provide. Imaging instruments have always fascinated me and I’m looking forward to witnessing how far we will be able to delve into the nanoscale world in the years to come, thanks to the fast pace of engineering and physics research. Previous article back to DISORDER Next article
- OmniSci Magazine
OmniSci Magazine is the University of Melbourne's science magazine, written by students. Read our recent issues and view the magnificent illustrations! Cover Art: Anabelle Dewi Saraswati READ NOW Welcome to OmniSci Magazine OmniSci Magazine is a student-led science magazine and social club at UniMelb. We are a group of students passionate about science communication and a platform for students to share their creativity. Read More More from OmniSci Magazine Previous Issues Illustration by Louise Cen READ ISSUE 6 National Science Week 'SCIENCE IS EVERYWHERE' PHOTO/ART COMPETITION VIEW SUBMISSIONS
- How to use a time machine | OmniSci Magazine
< Back to Issue 2 How to use a time machine Whilst time travel is thought to be nothing more than science fiction, the very laws of physics point to its possibility. Physicists have long sought the answer to such a phenomenon using knowledge from rockets to generating wormholes. by Sabine Elias 10 December 2021 Edited by Niesha Baker Illustrated by Quynh Anh Nguyen So you have just entered the TARDIS machine and are trying to work out how to use it to travel to the past to re-write the present and save the future? Well, look no further because you have come to right place. In this article, I will be describing how to jumpstart your time traveling vehicle and by the end, you will be proficient in navigating your way through the universe and evading time. Do be warned however, that batteries are not included and the simulation may crash at times. Now, you are probably wishing that you could travel back in time to have not clicked this article and saved yourself these two minutes of life that you will never get back. But is time travel really a possibility? We often think about the world as a state of order. Social and political constructs generally keep society running in a systematic manner. But what if I told you the entire universe came to exist from disorder? Before we get to logistics, let me introduce you to a little something known as ‘entropy’. Entropy describes the state of disorder (1). Take a closed bottle containing gas. Once you open this bottle, the gas will diffuse out into the open space with no way to retrieve it in the exact same state back inside the bottle. In essence, this gas has become ‘disordered’ and thus its entropy has increased. For years, scientists have understood that the entropy of the universe is always increasing, which means that stars, planets and galaxies are in constant motion away from each other (1). If we wanted to travel back in time, we would essentially have to reverse every single chemical reaction that has occurred from the point in time we currently stand in, to the point in time that we wish to travel to (2). This is theoretically impossible as we would be violating the laws of physics and decreasing the entropy of the universe but we still do not know if it is physically impossible. Let Brain Cox explain: Another problem with time travel would be altering events of the past. Take the Grandfather Paradox: if someone travelled back in time to kill their ancestor, then the possibility of their existence in the future would be zero (3). Thus, they would have been unable to time travel to begin with to have killed their ancestor. This issue of causality is expanded upon through the Novikov Self-Consistency Principle (4). This states that if an event causes a paradox or changes the past, the possibility of this event occurring would be impossible. However, this principle is not widely accepted by time travel enthusiasts. Now, whilst your TARDIS machine may be nothing but a prop at this point in time, it could still help provide evidence on the possibility of time travel. Take this example: you set up two duplicates of the same clock that read the same time and placed one into a rocket that blasts off into space. The rocket orbits around the Earth and then returns and is compared to the clock that remained on Earth. You would find that less time has passed on the clock that was in the rocket. Why? Because moving clocks run slower than stationary clocks. That is, as you move faster through space, you move slower through time. This is known as Time Dilation (5). An example of time dilation is the comparison of time on the International Space Station (ISS) to the time on Earth. Astronauts who have spent 6 months in the ISS have aged 0.005 seconds less than people on Earth (6). This does not seem like much because the astronauts are not traveling close to the speed of light. To see the effects of time dilation multiply, one would need to be very close to the speed of light. If you were to travel in space at 90 per cent the speed of light, whilst everyone on earth would age by 22 years you would only have aged by 9! Speed is not the only thing that affects how fast we age, gravity also affects our experience of time. A stronger gravitational field means that time travels slower in that field. For instance, your feet age slower than your head considering the slightly smaller gravitational pull on your feet compared to your head. Now take a black hole; we know that black holes have immensely strong gravitational fields where one hour near a black hole would equal approximately 100,000,000 years for a person on earth (7). So what would happen if you travelled through a black hole? No one really knows what occurs inside a black hole but we know trying to enter will likely turn you into spaghetti (8). That being said, we can only observe things that go as far as the event horizon of the black hole, so once something has entered it, we do not know what has happened. Black holes have however, been especially useful in theoretically explaining the possibility of time travel. Placing someone in a strong gravitational field or having them experience motions close to the speed of light would have them experience time slower compared to someone on Earth. This brings us to wormholes. Einstein’s theory of general relativity predicts the existence of wormholes which would theoretically permit time travel. To travel to a galaxy that is 2.5 million light years away with the fastest rocket on earth would be impossible as it would take longer than a human lifetime. This is where wormholes come to the rescue. A wormhole would provide us with a shortcut to our location of interest. Imagine folding a paper in half and poking a pen through it to represent your route of travel. You are essentially skipping the length of the paper and traveling from one end to the other. Source: The Independent. (2008). The Big Question: Is time travel possible, and is there any chance (9). You then situate one mouth of the wormhole in a spacecraft traveling close to the speed of light and the other mouth on Earth. If you then went through the mouth on Earth and travelled through to the space craft, you would be traveling back in time. This is because time would be passing much slower at the other end of the wormhole than where you entered from. However, physicists have not yet developed such advanced technology capable of this, but theoretically speaking, this is a possibility if such technology was developed in the future. Whilst you may have thought that time travel was merely based on science fiction, the laws of physics do not forbid its existence. However, here is some food for thought: “If time travel is possible, where are the tourists from the future?” Stephen Hawking Perhaps with time, we may transform this theory into reality. So for the time being, just sit back and enjoy the presence of your TARDIS machine. Perhaps you might even get lost in time from the very thought of time travel. References: 1. Wehrl, Alfred. “General Properties of Entropy.” Reviews of Modern Physics 50, no. 2 (April 1, 1978): 221–60. https://doi.org/10.1103/revmodphys.50.221. 2. BBC. “Brian Cox Explains Why Time Travels in One Direction - Wonders of the Universe - BBC Two.” YouTube, March 10, 2011. https://www.youtube.com/watch?v=uQSoaiubuA0. 3. Smith, Nicholas J.J. “Time Travel (Stanford Encyclopedia of Philosophy).” Stanford Encyclopedia of Philosophy, November 14, 2013. https://plato.stanford.edu/entries/time-travel/#GraPar. 4. Carlini, A., V.P. Frolov, M.B. Mensky, I.D. Novikov, and H.H. Soleng. “Time machines: The principle of self-consistency as a consequence of the principle of minimal action.” International Journal of Modern Physics, no. 05 (October 1995): 557–80. https://doi.org/10.1142/s0218271895000399. 5. The Editors of Encyclopaedia Britannica. “Time Dilation | Explanation, Examples, & Twin Paradox.” In Encyclopædia Britannica, 2019. https://www.britannica.com/science/time-dilation. 6. Dickerson, Kelly. “Here’s Why Astronauts Age Slower than the Rest of Us Here on Earth.” Business Insider Australia, August 20, 2015. https://www.businessinsider.com.au/do-astronauts-age-slower-than-people-on-earth-2015-8. 7. Gharat, Sarvesh Vikas. “Relativity and Time Dilation.” International Journal for Research in Applied Science and Engineering Technology 7, no. 11 (November 30, 2019): 650–51. https://doi.org/10.22214/ijraset.2019.11103. 8. "Death by spaghettification: Scientists record last moments of star devoured by black hole." NewsRx Health & Science, November 1, 2020, 236. Gale Academic OneFile. https://link.gale.com/apps/doc/A639405517/AONE?u=unimelb&sid=bookmark-AONE&xid=6812ee05. 9. “The Big Question: Is Time Travel Possible, and Is There Any Chance.” The Independent, February 8, 2008. https://www.independent.co.uk/news/science/big-question-time-travel-possible-and-there-any-chance-it-will-ever-take-place-779761.html. Previous article back to DISORDER Next article











