The HCS vLab is one of the Australian Government funded NeCTAR eResearch projects.
Human Communication Science (HCS) is a broad-reaching interdisciplinary mix of research that spans speech, language, sonics and music; and HCS research encompasses areas such as speech science, computer science, behavioural science, linguistics, music cognition and musicology, and sonics and acoustics.
The virtual lab:
- Brings together diverse data sets for the first time into one discovery/repository system, including text, audio and video resources with rich metadata.
- Links the data to tools – both inside and outside the lab, from simple bread-and-butter linguistic tools like concordance (keyword-in-context) searching and word frequency analysis to processing data using programming languages such as Python and R.
- Has a workflow system we borrow from the the bioinformatics community – Galaxy, where users will be able to chain-together complex processing pipelines on data.
Who might be interested? Anyone interested in text, audio, video processing of any kind will eventually have access to the vLab or be able to run their own version of it.
We’re proposing to give a demo and talk about how it might be useful to all kinds of digital humanities researchers, and seek suggestions about tools you might like to use in the future Forensic Literary analysis? OCR on pictures of graffiti? Tell us!
See more, including video demos at hcsvlab.org.au/.
OR – we could link up with other NeCTAR tools/labs at THAT Camp and do a joint session comparing our approaches.
Peter Sefton, Dominique Estival and Denis Burnham – University of Western Sydney
Data wranglers, back in the day, were spreadsheet wizards who could make any data sing to the tune of any report. Today we have data mashups magicians who can take data streamed from several places across the globe push them into Google gadgets displayed on any web page. They way we consume data today will undoubtedly change tomorrow. I have spent the last 5 years working with technology for a variety of humanities projects with a wide range of concepts, data modeling needs and data import requirements.
I would like to talk with researchers about their experience with data evolution throughout their research careers and the processes used to migrate data throughout the years. I hope that we get the opportunity to discuss approaches to evolving data across the decades, how to separate data models from work flows and why standards don’t last.
Paperboy icon, with the permission of the Artist Bradley David Santos www.bradleydavidsantos.net/works%5B/caption%5D
The concept is to be able to search for articles by institutions and find collaborators who are publishing in your searched interests. Three different potential user groups have been considered:
– Prospective students who are searching for institutions to study at
– Academics who want to find collaborators at other institutions.
– Policy and institution-makers who want to look strategically at the published record to inform decisions.
This experimental project was initated at ‘healthhack’ in Melbourne over the weekend.
What was achieved during the weekend is described on the wiki.
(N.B, the live demo is being updated this evening. Look at it on Tuesday morning).
Coders, Designers, Analysts want to make stuff during the un-ference and teach some others a little?
Digital humanities data is generally messy. Sometimes it looks like the floor of a teenage bedroom. Before we can analyse the data we need to clean it up and organise it. There are many ways of doing this. One important tool to use is a spreadsheet.
In this session I will run through some basic techniques in Excel that are often used for cleaning up data. This session should be useful for those who are starting out in digital humanities.
It would be good if someone who is proficient with OpenRefine (formerly known as Google Refine) could share this session so participants can get a taste for that tool as well.
I”ve put some simple Excel tips in a document which you can access here.