Neil Killeen (Research Platform Services, the University of Melbourne), Elena Pozzi (Orygen and the University of Melbourne) and Lianne Schmaal (Orygen and the University of Melbourne)
The Enhancing Neuroimaging Genetics through Meta-analysis (ENIGMA) Major Depressive Disorder (MDD) Consortium is an international consortium that pools brain imaging data worldwide to conduct large-scale studies aimed to identify patterns of brain alterations associated with depression, and to test their replicability and reliability across many different samples worldwide.
The first meta-analysis of the group on subcortical brain structures included data from 1728 MDD patients and 7199 healthy controls from 15 institutions (see Schmaal et al., 2016, ‘Molecular Psychiatry’ 1–7) and the second meta-analysis on cortical brain structures included >10,000 MRI scans from 20 institutions (Schmaal et al., 2017, Molecular Psychiatry). ENIGMA MDD now includes 35 institutions from 14 countries and 15 on-going projects . New research groups are continuously encouraged to join the consortium to increase our sample size and thereby the power of detecting meaningful findings. The ENIGMA MDD working group is led by Dr. Lianne Schmaal (Orygen and the University of Melbourne - chair) and Prof. Dick Veltman (VU University Medical Center, Amsterdam - co-chair), and is coordinated by Elena Pozzi (Orygen and the University of Melbourne).
Recently, the ENIGMA project team recognised that with the continued growth of the consortium, it was timely to enhance the way in which data were acquired, managed and distributed. Following discussion with Research Platform Services (ResPlat), it was concluded that one of ResPlat’s research data management (RDM) services (i.e. Mediaflux) could support the requirements regarding data contribution, sharing, retention, redundancy and security. There was no intent to change significantly the detailed way in which the data are held, merged and distributed (in spreadsheets), but more to provide a better environment to operate the study’s business processes (and to enhance them as appropriate).
The end-to-end workflow is that more than 30 external groups contribute data to the study into site-specific spaces. The ENIGMA Study team merges the new data into the master spreadsheet. Then, authorised subsets of the master spreadsheet are provided to end users (ENIGMA ‘projects’) by the ENIGMA Study team.

A key part of this workflow is to enable external user groups to securely upload their contributed data to a specific destination space, but ensure that they have no access to any other part of the system. Similarly, only the specific data for an end user ‘project’ must be securely accessible to the users of that project.
The main features of ResPlat’s RDM service that support this workflow are:
This study is a good example of how ResPlat’s RDM capabilities can be used in an agile way (a little development was required from Arcitecta, the vendor of Mediaflux) to support a bespoke workflow via generic and configurable Mediaflux frameworks. It also makes use of the unique multi-protocol capabilities of Mediaflux (in this case sFTP, SMB and HTTPS).
The only part that was slightly problematic is that a small number of the contributors work in clinical settings and their networks are heavily restricted. For the above solution, a hole in *their* firewall had to be opened to allow access to our system because the special sFTP service uses a non-standard port (the standard sFTP service on port 21 does not restrict users to configure home directories).
Since going live with this new system in July 2018, it has worked reliably and effectively. It has:
Senior Research Community Coordinator Errol Lloyd speaks with Bernie Pope, Victorian Health and Medical Research Fellow at Melbourne Bioinformatics, about his experiences analyzing large amounts of genomics data in cancer research.
What kind of a research world do you live in, and what do you care about?
It’s a good question. I care about human genomics, and human disease and cancer especially. I lead the human genomics and cancer group here and I’m funded to work on colorectal cancer. I’m especially interested in colorectal cancer and related diseases and I work very closely with the colorectal Oncogenomics group at University of Melbourne, who are situated in the VCCC building, so I spend a fair bit of time there.

The Melbourne Bioinformatics Human Genomics Group. Photo source.
How does your work with genomics relate to cancer research?
In the case of colorectal cancer, we spend a lot of time working on hereditary predisposition, so risk factors that might increase one’s risk for colorectal cancer. There are well known cancer syndromes that are familial, so we have aggregations, increased prevalence of cancer in a family far beyond what you’d expect by chance.
For example, a classic one is called Lynch syndrome. Lynch syndrome is caused by a defective DNA repair mechanism, so DNA gets damaged from time to time, especially during cell copying. When cells divide and produce new cells, the DNA is copied from the cell to the progeny and in the copying process some errors in mutations can happen. If those errors happen to a fall into important genes that regulate cell growth, that can lead to a cascade of mechanisms that might ultimately develop into a tumor.
We’re interested in finding the mutations that might be causing Lynch syndrome, because they can be tested for in screenings at cancer clinics. Other family members may also be advised to be screened for identified mutations as well, and advised on lifestyle changes and other things that can help reduce the risk of cancer developing in their life. Identifying cancer development at early stages or before cancer develops is the best outcome for patients. One thing we’re trying to do is clarify those cases like suspected Lynch syndrome, trying to work out more mutations that might be occurring. Some of them are more complex changes to the DNA, that are harder to detect with normal techniques at the moment.
So, that’s where you and research comes in, outside the clinic?
Yeah. I’m a computer scientist by background, so I have a computational angle on all of this but I collaborate very closely with biomedical researchers. We have to work together and genomics is really the tool for which a lot of this is done. It’s a technique that works very well and the cost has come down, so it’s becoming more and more widely applied. We’re able to do whole genome sequencing now, in many cases, which gives us most of the DNA in the cells, which is vast amounts of information requiring lots of storage on the computer but also a fair bit of computation, clever algorithms to deal with it to get answers out in reasonable amounts of time and so on.
On a day to day basis, where do you come in? What’s the magic that you lend to all of this?
What happens is that tissue samples are taken via other collaborators. Then the DNA sample is sent for sequencing and that can be done in many different ways. We might, depending on the study, sequence just a targeted region if we’re interested in specific genes. Alternatively we might do whole genome sequencing, which is look at all of the DNA in the cells and looking for mutations in that DNA. This would occur in suspected Lynch syndrome where we don’t know what the causes might be, and we might be looking for new mutations.
My role is to engage with that computational aspect, which is running some large computations to get from vast amounts of sequencing data down to biologically interpretable results and also develop novel techniques for annotating the mutation discoveries, so clever ways we can find new things. Also incorporating other datasets where possible, so we have a wealth of biological data out there. We can try to come up with interesting ways in which to use that information to improve our enrichment of truthful results. There’s a fair bit of data analytics and statistical analysis and so on that goes into it as well.
So what are the challenges you’ve come up against to get this work done?
Challenges around the computational side of thing are that we need fairly large amounts of storage. Over the many projects that I’m working on, there would easily be a few hundred, several hundred terabytes of data and it’s constantly growing, so that’s a pretty sizable amount of data to manage. Even just basic data management, it’s a fair bit of work. The data has cost lots of money to obtain in the first place if you think about all of the work that’s happened, so it’s a very precious resource.
There’s a pretty large amount of computation as well. It tends to be bursty, so when we get a larger amount of data and we do analysis, we do a very big calculation. That might take days or weeks but then there might be a period where we’re just then working with the downstream results, which I trivialized a little bit, but more or less looking at spreadsheets. There’s not a lot of computation going on there, while we’re we thinking about and analyzing that data, which you can do on a smaller computer. But then later on we might do a bigger calculation, so there’s these big bursts every now and then of large scale calculation, followed by intense analysis. I think there’s just practical, basic issues around getting enough storage, getting enough computation, being benefited greatly from being at Melbourne Bioinformatics. We’ve had lots of computing resources and also now using a lot at Research Platforms, especially Spartan. That’s growing and that’s great.
The more individuals we can see, the better. But we need ever increasingly large sample sizes because some of these mutations that we see in the DNA are extremely rare, say one in a few hundred thousand people. Just having enough samples is challenging, that’s globally challenging for everyone. People are aggregating their data as much as they can, so there’s people publishing public datasets and so on; it’s always growing but still a long way off the numbers we’d like to have.
More practically, it’s very noisy data. Processing the data to try and reach for a true signal is challenging, as DNA itself is a very complicated system. In humans and any cellular organism, the DNA itself in the cell is a very complicated system and so dealing with all of the complexity in there. It is quite a challenge from an intellectual point of view. We tend to simplify things down a little bit but the reality is it’s extremely complicated. Tumors themselves, depending on the cancer, can be very chaotic internally. They can be vastly different than what the normal cellular DNA looks like. For example, you can have extra copies of chromosomes, you can have total loss of other chromosomes, you can have neo-chromosomes, which are just joining two different ones together. The genome itself is a much more plastic thing then we think it might be.
Strange things can happen, you can have a big section of the genome inverted, just flips over and goes in the other direction. You have copies and inversions, you can have deletions and copies and inversions, you can have translocation where different things join together in unexpected ways and two genes which are not really connected to each other can break and then join to form and new gene that doesn’t previously exist, which has some effect on the growth of the cell. Looking at tumor DNA and genomes is a sort of interesting area and a very hard area to work in because it’s quite complicated. We do a fair bit of that as well, which I’m quite interested in.
Perhaps Researchers on campus are unaware of the infrastructure and support communities around them. Do you think it’s something Researchers require experts for using these resources or could they dive in themselves. What are your thoughts?
I think it’s very obvious that while the infrastructure is amazing and it’s very important, it’s not enough just to buy computing equipment and let people loose. It’s a difficult environment sometimes to use and requires specialist skills and so on. I think there’s a strong need for computational people, experts to bridge the gap. But not just to be an intermediary between the researcher and the computing equipment. I think actually my experience is to be a researcher as well, to add my knowledge and skills into the research plan, People bring skill-sets together and it adds up into being a much more powerful group.
You want people who can use the computers for sure, but also you want people who have that sort of research mind and understand what the collaborator’s trying to do.
The thing about accessing those resources that exist, the infrastructure’s already there, the key is just awareness, so knowing that what exists. And I guess organizations like Research Platforms and Melbourne Bioinformatics do some promotion as well. One of the main ways that happens is through training courses. And so we do heaps of training. You guys, research platforms do also, yourself do lots and lots of training. That’s a great way to introduce people to the ideas. And give them a sort of gentle introduction to things without jumping into the deep end.
While it is impressive that you have a room full of computers and they’re all grinding away, it’s rather more important to think about how your problem fits into how you’re going to solve it. And some problems are just very large, especially in biomedical sciences. But the practical side of things is not to be put off by the grandiose scale of stuff, just see it as a device that works for you.
Computational skills and knowledge is an increasing requirement across many disciplines, as more and more disciplines become more and more digitized and automated. And inevitably, researchers in many disciplines have got to pick up more of those skills and computational people will become more expert in other disciplines. Start asking questions like, where’s the research going? What questions can we solve? What technology do we need to solve those things? And think about writing grants around those things and so on. Think about collaborating with people, finding the right people who can provide the bits and pieces you can’t do yourself.
What do you wish you could tell your younger self? What would you tell your students here, or anyone who’s thinking of entering this world of biomedical research? What does the future hold for them?
That’s a good question. One piece of advice is not to pigeonhole yourself too much into some discipline. And in many ways, the world and academia tends to pigeonhole people. There are streams that you can follow, and when you follow those streams, you tend to kind of end up confined a little bit. And the modern world of research is very multi-disciplinary, it’s very collaborative, the idealized version of a scientist working in a lab on their own is not something I’ve experienced. It’s been large groups of people, working together over long periods of time. Research is much more dynamic than I imagined it would be so it changes direction a lot. Don’t be afraid to kind of cross disciplines, there are a lot of opportunities there.
You need to kind of follow where your interests lie. When you’re interested in something, you’ll naturally study it, you’ll naturally read about it and just think about it and so on. And that can’t help but improve your skills and knowledge in that area, rather than doing things that you feel other people are telling you are good things to do.
One more practical thing is I wish I’d studied more statistics because that’s a constant challenge for me and statistics is such an important part of lots of the work we do, so that’s something I’m always trying to improve upon.
To read more about Bernie Pope and his research, visit http://berniepope.id.au/

Associate Professor Martin Sevior. Photo: Eric Jong.
Martin Sevior is Associate Professor in the School of Physics at the University of Melbourne. As part of his research in the field of Experimental Particle Physics he performs experiments with the world’s highest intensity particle accelerator, Belle II at the SuperKEKB in Japan.
This experiment probes conditions that last existed less than 1 billionth of a second after the Big Bang and investigates the cause of the Universal Matter-Antimatter asymmetry.
It’s goal is to discover fundamental new physics not encompassed by the Standard Model of particle physics; what Professor Sevior says is generically called “Looking for new physics.”
Being particularly interested in the development of Machine Learning to make the best measurements possible with the Belle and Belle II data sets, Professor Sevior has been collaborating with Research Platform Services.
Research Community Coordinator Eric Jong sat down with him to talk about his research.
Could you give us a elevator pitch style overview of what you are researching?
So what I am interested in doing is making precise measurements of the standard model of physics, of processes that are well predicted of the standard model that nevertheless give different results to what the standard model predicts.
One place I am particularly interested in looking is in the phenomena called CP violation, which is essentially the difference between how matter and antimatter behaves.
Antimatter has exactly the same mass, and almost the same properties except for the opposite charge to matter. But we know that there’s an asymmetry because our universe is made of matter and not antimatter.
If you take the standard model and you put it in the model of the early universe and you run it all through, that gets the matter and antimatter asymmetry wrong by over 10 orders of magnitude.
So we know that there is some interesting new physics that is probed by making measurements in CP violation. And to do that I employ experiments called Belle and Belle II at an accelerator lab in Japan where we collide electrons and positrons.

Construction work on the Central Drift Chamber (CDC) of the Belle II experiment. Photo: Nanae Taniguchi.
So for your workflow, would it be right to say that you are making your data collections at Belle and Belle II in Japan, and then taking that data and processing it here at the University of Melbourne?
Some is processed on the world wide grid, on computers all around the world. Some is processed at the laboratory in Japan. And the final processing happens right here at the University of Melbourne.
So is that how you linked up with Research Platform Services?
Yes, kinda indirectly. The Centre of Excellence for Particle Physics employs two exceptionally talented computer professionals, Lucien Boland and Sean Crosby.
At one point my colleagues and I realised that we could make use of next generation machine learning technology. Which really needs powerful GPU’s to run. So the school of physics has a bequest, from which we requested funds to invest in one of these systems. We were able to get matching funds from other places and were able to get a few of these.
Sean Crosby was aware that Research Platform Services was putting together a few of these systems. They came to him and said, if you put yours in with ours - you can use all of them. So we did. We’ve been using the GPU systems that we initially purchased and also the GPU systems from Research Platform Services together in collaboration.
So it sounds like the machine learning for your research has been quite a cornerstone for the processing of your data. For a lay audience (such as myself) could you speak to that a little bit?
The problem with doing all of these measurements is distinguishing our interesting signal from a whole slew of random background noise. Our signal is less than one ten millionth of all of the data that’s actually collected. What we do with machine learning is make a important discrimination between those events where electrons and positrons collide and make processes that are interesting, and those that aren’t.
To do that we use machine learning techniques where we simulate the processes of interest. And we simulate the processes that aren’t of interest. Then we build a model that distinguishes the difference between the signal and the background, and then we train the model. It’s called classification. Every time there is a background we say this is a background and every time there is a signal we say this is a signal. And then the machine learning algorithm recognises what’s signal and what’s background and helps us make that distinction.

The classifier uses a neural net to combine many input variables to distinguish signal from background. The output of the neural net ranges between -1 and +1. Events near -1 are more likely background, events near +1 are more likely signal. By placing a threshold on the output of the classifier we can choose what fraction are signal and what are background. There is always a trade-off between signal efficiency and background contamination.
How would that have been achieved previous to machine learning? Was this process of classification something that you’ve gone through before using machine learning?
We’ve been using machine learning techniques in my experiments for well over 15 years, and probably longer, over 20 years. I like to tell people we’ve been doing data science since well before it was sexy. Or you could say, before it went mainstream.
So we have been riding the wave, and these new generation algorithms that use machine learning we’re still investigating. Because what we have now works very well, but we’re looking to see how we can do better using these modern techniques. And it’s possible, I think we can do better by at least a factor of two. Which helps enormously.
Do you have any advice for people in similar fields, or perhaps for people who are working with massive data sets, who are thinking about using a service like Research Platform Services?
First off, do the work. You really have to work to understand how it all works. Learn Linux. Learn how to use the command line. Learn how to do scripting.
Because all of this stuff that we do using large data sets involves taking one file from somewhere and putting it somewhere else, and processing it. And all of that requires some sort of algorithmic flow. There are techniques for doing that that are well established that aren’t what people who get trained with Microsoft products are used to.
So you really have to put in the hard yards to learn how to use them. And I am putting in the hard yards to learn how to use these modern algorithms too, it’s real work for me. It certainly hasn’t been easy, but a factor of two is a big deal to me.
If you have a project examining large data sets and are interested in learning Linux you can register for the next workshop at Research Platform Services here.

ResGrants is a new initiative that rewards Graduate Researchers at the University of Melbourne who want to teach a digital research tool.
KEY DATES:
Expressions of Interest open: 8th October
Information Session: 25th October, 4pm - 6pm
Applications due: 1st November
Winners notified: Friday 9th November
Winners present at the Research Bazaar Conference, University of Melbourne in February 2019
For questions or more information email: research.bazaar@gmail.com

A ResPitch explains the very basics of a digital research tool. It also:
- Engages the audience with a ‘challenge’
- Excites the audience about digital research tools
- Entices the audience to want to learn more
What are the 4 key ingredients of a ResPitch? …
Learning Objectives Sneak Peek
Briefly introduce self and tool. Understanding the origins of the tool, how the speaker was first introduced to and continues to use the tool can help in a number of ways:
a) increasing the validity of the tool
b) creates a personal touch by knowing who
created them and why
c) seeing how they are embedded in the broader
research community.
Learning Objectives might sound boring but are essential for learners to:
a) know
b) understand and
c) be able to do.
It is comforting to hear: “In the next half hour you will learn how to do A, B & C.” This focuses a participant's’ attention, points them in the right direction, and also gives them satisfaction at the end once they have achieved those learning objectives.
Give short demonstration or teaching section. It’s assumed there’s very little existing knowledge of each digital tool so knowledge will be slowly built up.
This concept is called scaffolding (Lev Vygotsky). By breaking large tasks into smaller, more manageable tasks, the audience can build on their current knowledge base to learn new concepts.
In the vein of scaffolded learning, provide the audience with a quick visualisation of the digital tool, orienting them to the environment they’ll be using in the challenge.
Each ResPitch has a challenge. These are based on the principles of Problem Based Learning. These techniques will ‘activate’ learners to engage.
At ResBaz we avoid participants having to open their laptops and downloading tools – this takes away precious time from your pitch. Find three or more helpers with the tool already downloaded on their laptops. Assign participants to small teams of three or four with the helpers at the helm leading the challenge.
Learning-by-doing is a powerful tool. You will never understand the pitfalls better than if you have already made the mistakes yourself.
Each ResPitch ends with a plenary that:
…Lastly, have fun during your ResPitch and enjoy connecting with like-minded people!

Ashton Dickerson, Biosciences PhD researcher and member of the Urban Light Lab. Photo: Eric Jong
Working smarter with Ashton Dickerson
Using High Performance Computing for increasing efficiency of research.
For the last year Ashton has been using Spartan with a PhD project that examine the effect of light on the nocturnal songrate of Willie Wagtails.
By using a automatic song detection package through R to extract data from the over 2000 hours of audio recordings she has gathered in her field work, Ashton has been able to automate the otherwise labour intensive handling of this data.
Then by working with Research Platform Services, Ashton has been able to complete these processes on a HPC system where large numbers of these tasks can be run simultaneously, saving her time that she can use of other aspects of her research.
Research Community Coordinator Eric Jong sat down with Ashton to talk about her project, and how she is integrating high performance computing into her workflow.

A Willie Wagtail. Photo: Timon van Asten.
Can we start with the question that I’m sure you’ve answered a million times now as a graduate researcher, what are you doing your PhD on?
Well, I research a quite unusual behaviour of birds that not only sing during the day but at night time as well.
Some diurnal (active during the daytime) bird species, also sing during the night time. This is an unusual because you would instead expect these birds to be sleeping during the night.
For my PhD I aim to understand why diurnal species are singing during the night.
To answer this question I have been examining this behaviour in an iconic Australia species, the willie wagtail (Rhipidura leucophrys), who has a reputation for its prolific nocturnal song.
So it sounds like a big part of your PhD is listening to the song of the Willie Wagtail, how have you been gathering this data so far?
To measure nocturnal song, I use bioacoustics recorders from Frontier Labs that allow me to record audio for prolonged periods. I target the roosting spots of willie wagtails to record their nocturnal song.
Thus far I have gathered over 2000 hours of audio.

A researcher checking a bioacoustic recorder. Photo: Justine E. Hausheer / TNC.
That is a huge amount of data, can you talk a bit about how you have been handling that volume of information for your research?
To be able to handle such large data sets I am utilizing an R package, monitoR, which automatically detects bird song.
I import templates of willie wagtail songs into this package, which then is run along my recordings and it detects when the template matches a song. From this data I can extract the song rates (how often the willie wagtails are singing) and then I can examine the data to look for patterns.

A spectrogram showing an example of the automatic song detections from an hour-long recording. Blue line indicates where the R package, monitoR, has detected a willie wagtail. Image courtesy Ashton Dickerson.

A spectrogram showing an example of the automatic song detections from an eight minute long recording. Blue boxes indicate where the R package, monitoR, has detected a willie wagtail. Image courtesy Ashton Dickerson.
I am using the Spartan service through HPC at the University of Melbourne to be able to handle such large data loads. Lev Lafayette has assisted me by uploading my audio recordings to the UniMelb cloud, which is much more efficient that uploading this data via my personal computer.
The HPC is significantly faster than running these scripts on my personal computer. It would take me about 7 mins to process 1 hour of audio this way. Now using HPC it is about 3 to 4 times faster.
And in addition I am able to run this script over multiple recording sets at one time thanks to the multiple nodes. Not only does this save me immense amounts of time, this also means my personal computer is free for me to use while this data is being processed.
One of our mottos at Research Platform Services is ‘work smarter not harder’, which I think you are most definitely doing by automating these processes. Do you think there are things that you are able to spend more time with now in your research because of this?
Most definitely so, it frees me up to read papers and continue researching. To form thoughts and ideas around what this data actually means.

Ashton using HPC to give her more time to do MAXIMUM SCIENCE. Photo: Eric Jong.
Using HPC has allowed me to take away the manual processing and gives me time to think about what this data actually means, to analyse it and put together ideas from it.
Thus far from the data I have extracted using the HPC services, I have discovered that willie wagtails’ nocturnal song significantly increases with lunar illumination, showing that this behaviour has a relationship with light, and therefore may be related to a visual cue.
This is an interesting finding and gives insight into the possible function of nocturnal song, furthering our understanding of the evolution and function of bird song in general. I am now preparing a manuscript for this finding.
Furthermore, given that I have discovered that nocturnal song has a relationship with light, I will also examine if this behaviour also responds to artificial light at night (e.g. streetlighting), which could highlight a possible stressor for urban bird species with nocturnal song. I will again utilize the HPC services at UniMelb for data extraction.
Thanks for your time today Ashton, do you have any advice to share with other researchers?
The first step for me using HPC was just hearing whispers that something like this was possible, and then from there I looked for and found people who could help me with it and point me in the right direction, and also gave me different options to choose from.
So I guess I would say it’s all about building and engaging others in the research community.
Visit Research Platform Services for more information on HPC and other services.
Congratulations to Sara Ogston and Koula Tsiaplias for their work in Vic ICT for Women’s 2018 Grad Girl program which won two TechDiversity Awards on Thursday the 27th of September!

Sara Ogston and Koula Tsiaplias receiving the award.
The Vic ICT For Women Grad Girl 2018 program was recognised in last week’s TechDiversity Awards as a valuable contributor to diversity and inclusion in the STEM sector, for which it received two awards;
This achievement was celebrated and shared with the sponsors of Grad Girl, one of them being The University of Melbourne, for their support and commitment to making a difference, and with the grad girls, who committed in this year-long program on top of their academic and work schedules.

GRAD GIRLS IS A 1 YEAR PROGRAM RUN BY VIC ICT FOR WOMEN FOR FEMALE UNDERGRADUATE STUDENTS TO DISCOVER AND UNDERSTAND THE PATHWAYS AVAILABLE WHEN TAKING THE NEXT STEP IN THEIR CAREER.
Read more about the Grad Girls program here.
By Emilie Walsh
Alison’s experimental film using Fusion 360
One of the exciting part of working for Research Platform services as a CAD and 3D printing ResCom*, is to get to meet researchers working with 3D in all disciplines. Alison Kennedy is currently a Master by Research candidate at the VCA, and has been coming to our trainings for a few months now.

Alison Kennedy, Self-portrait, 2018
What I find fascinating
with researchers in Visual Arts, is how they take ownership of a digital tools,
push the limits of the applications sand find creativity in the often
frustrating glitches and bugs
I’ve asked Alison to tell me about her use of CAD and 3D scanning in her research and art practices. It’s intriguing to see how engineering, archaeologists, designers, and artists use CAD in very different ways!
Emilie: Alison, tell us a bit about your art practice?
I am particularly interested in how art gives artists a platform for commenting on and taking a position in relation to things happening in the world. I think that art can provide a way of suggesting a response without being didactic. For me this is because art, once created, allows the viewer to complete the artwork through their own personal experience. This tension between what is intended by the creator and what actually occurs is a constant fascination to me. My use of technology arose out of this - I started creating a series of collages and digital paintings that used and were generated from collapse, breakdown and error. These glitches represents the slippage between intention and creation and the uncovering of personal truth. We are both furthest away and closest to ourselves.

“Untitled: Force of Reason” 2016 120cmx120cm digital painting/ collage limited edition giclee print.
My initial work in technology concentrated on digital painting and referred to romantic narrative paintings of the 18th century. I wanted to reconsider the human gesture - how embodied expression translated through the medium of the mouse, and stylus. I started to consider how texture and colour transformed completely through algorithmic extrapolation and started to use this quality in an intuitive way to express personal environmental concerns.

“Untitled” 2016 20x24 cm limited edition giclee print.
Emilie: How has 3D scanning and modelling have bring new direction to your work?
I became aware of the potential of 3d technology applications to express the body in a totally new and unusual way that I believe critiques our approach to other people and to the world. I am interested in taking existing applications, hacking into them and pushing them to the point of collapse - at this point I think that something new and quite profound occurs. Again, I work with technology intuitively and at this stage in my research I think that the constant creation and destruction inherent in the process highlights the relationship with the world in general.

Still from animation “Selfie” https://alison-kennedy-gdg8.squarespace.com/config/
In the work above, for example, I wanted to show how the artist in her studio can make a stand in relation to the world and I also wanted to suggest that at times the artist’s studio is a claustrophobic space. Personally, I love working in my studio so in a way this was quite a confronting idea for me. The figure ultimately breaks down through algorithim and is revealed as a series of surfaces - which is an idea I’m really interested in and working with in my research.

Still from an experiment in Fusion 360
At this stage I’m most interested in how 3d packages critique image and our image saturated society. The packages I was introduced to at Research Platforms at University of Melbourne connect engineering CAD and create surfaces and objects. Once again I am interested in how new approaches to these standardised applications expose how technology and our world interact.
If you are interested in learning to use CAD, 3D modelling, scanning or printing for your research, be in touch with us at Research Platform Services!
by Emilie Walsh
For my last day working for Research Platform Services, I wanted to organise a meet up at the VCA where I did my PhD. The campus is on Southbank, and sometimes researchers there find themselves a bit far away from the services that are offered at Parkville, so we make sure we offer some of our training in the other university campuses.

At the digital lab at the VCA
3D modelling, 3D scanning and 3D printing can be an amazing tool and resources for researchers at the VCA, in Fine Arts, Music, Theater or Dance Studies. Of course we invited researcher from all discipline too, as we believe in the emulation it creates and we always get excited about researchers collaborating across disciplines!
Our meet ups at Research Platforms are not your usual tech training: it’s more about sharing research projects and talk about the digital skills we use as researchers
First we talked about the benefit of 3D scanning to share and collaborate. If you are working with fragile artefact, you may not be able to access it, manipulate it, or share it with other researchers. Having a digital 3D model is a great archive and tool to share your research.
Drag and drop script: 3.3M vertices > 20K vertices + normal + AO + displacement + centre geometry. No material editing required in Sketchfab 😀 https://t.co/OZhAfAEpoh pic.twitter.com/S7zgGC6k07
— Ben Kreunen (@OzBigBen)9 September 2018
You can also 3D print a replica for teaching or communication purposes.

A 3D printed replica of a skull
If you are interested in working with object based data sets, and use 3D modelling and 3D printing, you can read more about it here
Then we welcomed Tall Ben, from the digitisation services of the university.
He presented the 3D scanning technology available for researchers. If you need a 3D model of a tiny insect or of a large architecture building, Ben is the guy for you!
Our workstation is getting old now but still fast enough with @RealityCapture_ to make a draft while shooting, just for peace of mind. It’s the weekend now TBC… pic.twitter.com/XYaZqJ0rkN
— Ben Kreunen (@OzBigBen)31 August 2018
Next, Mitchell Harrop, from the digital lab in Arts West in Parkville presented some projects the Digital Studio in is supporting. 3D modelling can also allow you to archive architectural artefacts, geo-localise them, and embed photos and other documentations to display online and communicate research better.

More about Mitchell’s project :
https://people.eng.unimelb.edu.au/mharrop/mhw/v2/
Eric Jong, Master student and the VCA and ResCom at ResPlat, shared his latest 3D printed experiment.

Screenshot of Eric’s early experiment with a 3D model of a soundwave, in Fusion 360
3D printing is a cheap and fast technology, that allow makers to prototype, fail fast and made adjustment much quicker than with traditional technologies.
https://www.instagram.com/p/Bl4tpeOFyau/?utm_source=ig_web_copy_link
To end the meetup, we did a quick fun demo: How to do a 3D scan with your phone!
We 3D scan a cheese board in a few minutes for some amazing result
There are of course a range of options if you need a 3D scan for your research: the digitisation services at Uni would be the highest quality, but will require a lot of time, collaboration, will generate a lot of data, and potentially have a cost. A 3D scan with your phone would be the quickest options, with a lower quality in a DIY spirit!
If you are interest in 3D scanning some of your object based data set, be in touch with us at Research Platform and we can point you to the right option for you!

The whole team of researchers at the digital lab, VCA
Meetups are a great way to meet other researchers using similar tools than you, and work on solving some of your problems together. We alternate between meetups and trainings. If you are interested in joining a training check our calendar or be in touch with Eric for CAD and 3D printing.
by Emilie Walsh





If you looking for a hard copy of that little comic on 3D printing, come over at Colab to pick up one, and join a training in 3D modelling and 3D printing with Eric Jong!
Next training the 10th of October (cake included) :
https://www.eventbrite.com.au/e/introduction-to-3d-printing-with-tinkercad-tickets-50681449580
Have you ever wanted to type some mathematics in a document?
During a meet-up on the 5th of September, 2018 we discussed how to typeset mathematics, using LaTeX and JavaScript. Keep reading to learn how, and to view some beautiful examples (including animated rainbows!)…
Do you want to use mathematics in your documents? Join us tomorrow afternoon to learn how, using LaTeX and JavaScript!
Register for free: https://t.co/2w4Ec1CnMS
…see examples, learn tips & tricks, and ask questions while enjoying light snacks and beverages with the community! pic.twitter.com/YIv2nzVN1C— Meirian (@MeirianLT)September 4, 2018
LaTeX is a programming language for high-quality, beautiful typesetting. TeX was originally released by Donald Knuth in 1978. As a mathematician himself, Knuth was motivated to create something which allowed one to type complicated mathematical formulas in a professional-looking manner. As a result of his efforts, the ability to typeset mathematics in LaTeX is a major strength.
Anyone who has studied mathematics or statistics will know that the notation is plentiful. LaTeX allows for the many symbols you may desire to use, whilst presenting them in a clear and beautiful way. So many symbols are available that there is a 338 page Comprehensive LaTeX Symbol List. But when it comes to finding the symbols you need, I recommend using Detexify by Kirelabs. This tool allows you to simply draw the symbol you’re looking for, then it will identify it and tell you which command and package to use!
During the meet-up we created a LaTeX document which was abundant in examples of mathematics notation and formulas. You can view the document on Overleaf.
Once LaTeX had been introduced, Errol then gave an excellent presentation on the inclusion of mathematics on the web. Read on to learn more and see his animated rainbow equation!
Thanks to @maegul for your presentation on #JavaScript during our meet-up today on Mathematics in your Documents at @ResPlat! Your use of emoticons, rainbows and animation captured everyone’s attention! pic.twitter.com/h5B33B7ZdM
— Meirian (@MeirianLT)September 5, 2018
One of the great things LaTeX has given to the world is its mathematics or equation typesetting syntax. Not just because LaTeX is great and powerful, but, because this typesetting syntax is now a standard across many different kind of software.
This is especially the case when it comes to web pages. Go to the wikipedia page on the normal distribution, and see all of the mathematics typesetting. Copy one of the equations and paste it into a text file … what do you see? (Spoiler alert … LaTeX).
All of this works because when it comes to the programming language called javascript, which powers all of the interactivity and animation of the web, there’s a principle which has shown to be true again and again: ‘anything that can be made in javascript will be made in javascript’. The web is everywhere and everyone uses it all of the time. Which means javascript has to do its best to be everything to everyone. And so, LaTeX mathematics can be done with javascript.
If ever you find yourself wanting to write a blog, create a simple web page (which is easier than you might think), or share your Jupyter or RMarkdown notebook or create an attractive interactive data visualisation or publication for your research, know that javascript has got your back when you need to show some mathematics.
Beyond blog writing, you may be wondering why you should be interested in javascript and making web pages. There are two responses to this. First, with the Observable Notebook, it is now easy to learn and get started. Check out the tutorials at the home page and just start playing. Second, unlike any programming language you’ve learnt, javascript is built to normal programming things like crunching numbers, as well as graphics. This means you can make interactive publications, like this notebook on predator and prey population dynamics.
Or, you can just make rainbow equations:
#javascript + LaTeX = Awesomeness @observablehq https://t.co/V7KTWx3IlL pic.twitter.com/b1VfXhJxXV
— Errol Lloyd (@maegul)September 6, 2018
If you’re interested to learn more, take a look at this notebook of mine, which demonstrates LaTeX, interactivity, using Observable, and making interactive graphics with javascript.
This blogpost was created by Meirian Lovelace-Tozer (@MeirianLT), and Errol Lloyd (@maegul) who are Research Community Co-ordinator and trainers at Research Platforms Services (@ResPlat).
Why would you want to learn Numpy? Well I guess I can start by telling you why I learned it!
I don’t often admit this, But I started as a Matlab user (gasp!). Needless to say it is an excellent product, but sadly in many professions, it is quite a bit too expensive. So I found myself looking for an open source replacement. Also, I wanted to learn Python because it got along well with the GIS programs I often use.
However there is a bit of a learning curve: In Matlab you can easily do vectorized mathematics or matrix mathematics on two vectors of data like so:

And that is hella convenient! The maths look like maths!!!
Now, when I first started using Python, I was told that it was basically like Matlab so I tried to do some basic vectorized operations:

That is not what I wanted!!
Python will actually concatenate, or splice, the two lists together into one list, instead of adding up the items. This is because lists can store any type of data or python object, so Python wouldn’t necessarily know how to add up each item in the list like we saw up there… hmmm A bit frustrating!
Thank you for your realism Borimir. Now a Python Purist would use a list comprehension, which is basically a quick one liner loop creating a new list

**Opinion: **There is a lot of riff raff around the math up above, which makes it a bit harder to read and comprehend, plus, if we accidentally have a non number in our inputs, python will flag an error right?
Well luckily people wayyyy smarter than I have addressed my concerns, and have created the Numpy Library in order to make maths look like maths again in Python:

As you can see above, we can turn any list we want into a numpy array for analysis, it just follows one major rule:
Numpy Array Rule 0 - All data in it are coerced into the same data type,
that way python knows how to add the elements together (and do other mathematical operations)
But wait there is more!
Because of the above rule, Numpy arrays are much more efficient than lists at storing large data sets of numerical data, and cycling through them than a list
Numpy arrays can be multiple dimensions, representing geographic data, 3-D data, and any sort of multidimensional data set
….AND Numpy arrays are used to feed data into a variety of other data science packages like Pandas, Scipy, Scikit-learn and plotting libraries
So how do I Numpy?
Well I am a Geomorphologist, which is a fancy term for an Earth Scientist who spaces out so much on car rides they wonder why that hill over there is a particular shape.
#realgeomorphologyquestions - answer - It’s very pointy
I literally study the shape of the Earth’s surface. In many cases, the data describing the earth’s surface, and the factors changing it, can be described as a “raster” (basically photograph with pixels)
where we lay an imaginary grid on the earth’s surface to separate it into even sized squares, and store a value (say elevation) representing the value averaged over that square in the grid. Here is some 2 metre by 2 metre resolution LiDAR elevation data as an example:

Note, there is a bridge crossing the river from north to south in the middle,see if you can spot the road fill! The darker red colors represent lower elevations, and therefore old (and current) river channels. I could stare at this stuff all day. That’s why we in the biz call it LiDAR crack… because it is physically addicting to a Geomorphologist!
And with a 2 dimensional Numpy array, i get a lean efficient fast way to to analyse my LiDAR data using open source software, and figure out why that river switched places up above!
But wait there’s more!
if I am feeling really fancy, I can use a third dimension representing time slices of lets say a flood simulation, and make a movie of the water velocities (note matplotlib helped here as well)
SO Are you ready to calculate smarter not harder? Then check out our course material here come on down to our numpy workshops (see calendar) and learn how to crunch the numbers with Python
Cheers,
Jon
At the recent Omeka Meet-Up, we opened up discussion to include all things Digital Humanities.
Over three fantastic presentations and several provocations around data, Humanities research and cultural collections, the group developed a range of responses, provocations and ideas for future projects.

To kick start things we first had a discussion about what we mean when we use the term ‘data’ in the Digital Humanities space. Are we referring to text? Images? Digitised materials? Or simply everything involved in research? Questions around ontologies and silenced voices also came to the fore as did thoughts about the future of infrastructure in Digital HASS.
Julianne Bell then gave a fantastic presentation on a project she is working on as part of the Digital Studio Graduate Internship program, entitled Execution Ballads of Pre-Modern Europe. Developing this project with Dr Una McIlvenna from the School of Historical and Philosophical Studies, Julianne has encountered an array of challenges and discoveries around presenting and narrating cultural materials with Omeka.
The project has developed a database on execution ballads (cheaply printed songs about crime and punishment dating from c. 1550-1900) that includes hundreds, if not thousands of items. These are partly in textual form (transcriptions of lyrics, notes on historical context, dates of publishing, etc), partly pictorial (images of the pamphlets and broadsides, images of historical figures who feature in the ballads, etc), and some audio recordings. Until recently, this has been stored in a MySQL database on Heurist software. The intern project is transferring and setting up the database correctly in the content management software Omeka, as well as adding to the database.
Here’s a breakdown of the project:

And the methods and approaches taken in customising Omeka to fit the parameters and requirements of the project:

Following Julianne, our excellent Omeka ResLead Alex Shermon spoke about his work on an Omeka site that is part of a broader project within the University of Melbourne looking at Sir Redmond Barry.
This project brings together cultural and legal history, biography, philosophy and a wide range of curation and collecting practices.
Here’s the official description from the Lives Lived with Law Journal edition:
“We see ‘Lives Lived with Law’ as drawing into relation the scholarly experiences of disciplinary technique, and the experimentation over time with style and forms that help to show what the conduct of lawful relations can be between peoples, between everyday and official experience of law, as well as between Indigenous and Anglo Australian laws.” (Genovese, Rush, McVeigh. Lives Lived with Law: An Introduction, p.2)
According to Alex, “Jurisography is the hip new rebranding of legal biography.”
Working under Carole Hinchcliff (who is collaborating with several academics from the Melbourne law school), Alex has set out to examine the fragmentary sources and forms of legal theory involved with Sir Redmond Barry.

Title: The trial of Ned Kelly
Subtitle: Newspaper illustration
Date: 1880
Keywords: people, biography, Bushranger, trial, illustration, media
Record creator: Department of Information
National Archives of Australia
Alex writes how, “Sir Redmond Barry (who, by the way, has an apple named after him) got a lot done and left a pretty big legacy. Aside from the apple thing, he helped found the State Library, The National Gallery, and the University of Melbourne (He would have loved the sheer boring monotony evoked in the brown brick monstrosity named after him).”
“In law he was the first standing counsel for Aboriginal people, and famously sentenced Ned Kelly to death just days before he himself died. Barry shattered over Melbourne when he died. The task then for the Jurisographer, is to pick up these pieces and present them in a meaningful way. That’s where Omeka came into our project. It is allowing us to relate contemporary instances of Barry (statues and bookcases) with his own digitised publications, such as the important addresses he made at circuit courts and his annotations in statute volumes.”
Alex presenting his talk:

Mitchell Harrop from SCIP then delivered some timely and useful information about Omeka and Web Archiving - a topic we’ll be coming back to very soon as there is increased need for and interest in researchers being able to maintain and present their work on Omeka into the future.
Stay tuned for the next Omeka event. We have something very exciting in the pipeline which will, once again, consider dynamic scholarly workflows that connect Omeka with other digital research tools.
For more information, get in touch with our Community leader, Tyne!

Humanities, Arts and Social Sciences researchers are often faced with big questions when it comes to planning and orchestrating their data-related projects.
Our recent Omeka workshop took some of these questions and challenges head-on by considering the research ‘life cycle’ that comes before and after using a web-publishing platform.
While Omeka is free, flexible and very easy to use, there are serious and important considerations that must be undertaken before embarking on a research project that has a visual exhibition component.Full house for @omeka workshop today @ @ResPlat with 2 excellent ResLeads, Alex & Daniel! pic.twitter.com/OEGCcZT0MR
— Tyne Daile Sumner (@tynedaile)July 26, 2018
What kind of story do you want to tell?
Who is your audience?
How will the project or site be used?
What kinds of items will be included in your project?
What kinds of features do you want? Will they enhance the research?
Who will be contributing to the project?
Is the project sustainable?
In the workshop we explored provocations such as these. We trialled this new approach to gain insights into the kinds of questions that HASS researchers ask at the outset of their projects as well as some of the barriers that they face and how these might be overcome. The result was an engaging and thought-provoking discussion around not just Omeka but Digital Humanities research outputs more generally.
Jared Winton opened the session by canvasing some of the key recommended steps when starting out in a HASS research project that involves collections and data. Jared is the Data Collections Manager at Research Platform Services. ResPlat offer a range of research-specific infrastructure. In the workshop we had a brief look at Mediaflux - just one tool among many for managing digital assets.
We then looked at Omeka in the context of a data management plan and considered how best to set up a project. Although Omeka is designed with non-IT specialists in mind, it’s nevertheless important to seek out expert help in the early stages of a large research project.

Omeka can handle large archives of metadata and files (over 100,000 items), and is extensible by publicly available plugins and themes. Users can populate an Omeka site by adding items individually or by ‘batch,’ using data migration tools such as the OAI-PMH harvester, and CSV, EAD, and Zotero importer plugins. Omeka accepts and stores all file types, including: images, video, audio, multi-page documents and Pdfs, and Powerpoint presentations.
Our fantastic ResLead Daniel then gave a demonstration of just one way in which a large dataset can be imported into Omeka. Daniel’s example was the Actors Database on IMBD, which he used as a case study dataset.
Using a Python tool (Pandas) to wrangle the data into shape, Daniel showed us the steps that could be taken to transform a pre-existing database for display in Omeka.And here’s a shot of today’s guest presentation by Daniel on preparing data for @omeka. @ResPlat pic.twitter.com/whcasijEql
— Tyne Daile Sumner (@tynedaile)July 26, 2018
Finally, one of the most important parts of conducting research that uses Digital Humanities methods or tools, is being able to connect with the relevant experts.
Here at Melbourne University, two of these resources are:
Research Integrity Training (RIOT): covers the principles of research integrity and their application to the planning, conducting and reporting of research.
Managing Data @ Melbourne: an online program that guides you through the process of writing a data management plan for your research project. It has been developed for graduate researchers, early career researchers, and information professionals.
There’s lots going on in the world of Omeka at the moment.
We’re working on making “Preservation and portability” better via #DataCrate. Eg, this export from Omeka: https://t.co/t6OVtuKMKV @omeka https://t.co/Yd8ce7l3jI
— petie (@ptsefton)July 30, 2018
On the 3rd of August, 2018 we held our first LaTeX meet-up!
The purpose of this event was not only to showcase some exemplary LaTeX documents, but also to give attendees the opportunity to ask their burning LaTeX-related questions, and learn some tips and tricks. We saw examples of including images in LaTeX, a recipe book, academic posters, and slide presentations. Keep reading to hear about each of these, and to access some great templates…
Do you use LaTeX? Come along and meet other LaTeX enthusiasts tomorrow afternoon for a special event! https://t.co/mvtMQ5sLDZ
…ask questions, learn tips & tricks, and see some exemplary documents while enjoying light snacks and beverages with the LaTeX community! pic.twitter.com/d0gLBm8dN5— Meirian (@MeirianLT)August 2, 2018
After some quick introductions and once everyone had settled in with some tea and snacks, Ben began the meet-up with his presentation on preparing illustrations and images for inclusion in a thesis using LaTeX.
Ben also discussed the following frequently asked questions:LaTeX handles images slightly differently to many other applications. The DPI setting in the image is ignored and the size of the image will be based on the DPI of the document if no other size parameters are included. There are no rules for adding images, thesis instructions including required DPI settings are rare. For the sake of consistency I recommend you resize your images to match the size that they will be reproduced at… There are a few ways of including figures with vector objects (with or without images as well) in LaTeX. From a technical support perspective I look for methods that provide the best quality with the fewest problems… – Ben Kreunen @OzBigBen
Next, Brian presented his Mother’s recipes coded in LaTeX. The detail of his presentation was not about code specifically, but about communication with audiences and working with teams.
When we write we often write to ourselves, as my Mother did. We may include shorthand, not fully explain ideas presuming others have already grasped them and more importantly, we may not be actually addressing our audience or have a clear idea who they are. Pitch your ideas at the knowledge set of your presumed audience and explain all concepts that would be unclear to this audience. Make sure the flow of the document is logical and leads the reader to the conclusions that you intend.
Also, working with teams on documents — as this one was — has it’s own difficulties and complexities. This was created with an online editor/UX coder, a typesetter/graphic layout person and a recipe checker. As it was a labor of love it was difficult to assign deadlines. However, when working with a team who have specified areas and content to deliver it is a good thing to specify both job demarcations and time-lines in which to deliver the work. Deadlines and clarity around what a person is to do helps them deliver and, while it may be some pressure, people generally perform better knowing the limitations. – Brian Boyce @innerversitysou
You can view the template Brian created for the cookbook on Overleaf.
Fantastic #latex meetup tonight with three awesome speakers @OzBigBen, @innerversitysou and Tom, each sharing their experiences with the tool. #resplat #unimelb #research pic.twitter.com/WGrVJfTmWI
— Research Platform Services (@ResPlat)August 3, 2018
Meirian also showed examples of slide presentations which were created in LaTeX.These included his Ph.D. thesis, a presentation, as well as a set of lecture notes for a subject he is teaching. Particular attention was given to creating a poster in LaTeX using the baposter template designed by Brian Amberg.
This link contains a sample poster created using the template, the documentation of the template, as well as a LaTeX primer with sample code to get things rolling. – Thomas Wong
The document class for creating presentations is beamer. There are many options for layouts and colour schemes to choose from; some of these are displayed in the beamer theme matrix. I recommend reading the ShareLaTeX guide on beamer for some great help and tips when creating your own presentations in LaTeX! There are also a variety of templates available online to help you get started. – Meirian Lovelace-Tozer @MeirianLT
Would you like to come along to a future LaTeX workshop or meet-up, and get involved?
I encourage you to sign up to upcoming events on Eventbrite. Please also check out our website for more information, or contact me if you have any questions.
It is time for us to grow our ResFamily again! Do you love science and research? Are you the one all your friends/colleagues come to for their typesetting solutions? Can you explain technology eloquently and kindly to the grumpiest professor? If you answered YES to any/all of the above, then Research Platform Services invites you to apply for our Junior Research Community Co-ordinator position in 3D Printing and TinkerCAD!
The successful applicant will grow their respective communities through regular workshops and meetups. Maintaining an online presence is also an essential part of the job. You would also be expected and supported in organising events within your communities, as well as Research Platforms-wide events such as the famous Research Bazaar conference. You can find out more about the Research Bazaar community in the first pages of our new publication: The Digital Research Skills Cookbook.
Please see the below links for a detailed position description for the role, including how to apply. Get in quick, applications close COB 16th July, 2018.
3D Printing & TinkerCAD Position Description
