Survey Research Methods Webinars Spring 2018

The Survey Research Laboratory is offering two webinars on survey research methodology during the Spring 2018 semester. The webinars are free to University faculty, staff and students. All webinars begin at 12:00 p.m.

ADVANCE ONLINE REGISTRATION IS REQUIRED

You will receive a reminder about the webinar for which you have registered shortly before the date. Webinar notes will be available here shortly before the webinar.

Introduction to Survey Sampling
Wednesday, February 14, noon

Presenter: Linda Owens

This webinar will cover the basics of sampling methodology: the importance of using proper sampling techniques, determining the appropriate sampling methodology, and calculating necessary sample sizes. The discussion also will include simple random sampling, cluster sampling, stratified sampling, and multistage samples.

Introduction to Questionnaire Design
Wednesday, February 21, noon

Presenter: Allyson Holbrook

Designing a good questionnaire is a complicated process that includes decisions ranging from questionnaire format and question order to question wording and response categories. The design should aid respondent understanding of questions, recall, and judgment formation, and minimize response editing because of social desirability. This workshop will review basic strategies for achieving these goals.

Survey Response Rates: Uses and Misuses
Wednesday, February 28, noon

Presenter: Timothy Johnson

What is a “good” response rate, and why does it matter? These are common questions that we see at the Survey Research Laboratory. This webinar will provide a basic overview of survey response, cooperation and refusal rates, their uses, and why they are often imperfect indicators of survey quality and representativeness.

What Storify Shutting Down Means to Us

The Storify logo.

You may have heard that popular social media story platform Storify will be shutting down on May 16, 2018. Open to the public since 2011, it has hosted everything from academic conference tweet round-ups to “Dear David”, the ongoing saga of Buzzfeed writer Adam Ellis and the ghost that haunts his apartment. So it shocked long-time users in December when Storify suddenly announced that it would be shutting down in just a few months.

Already, Storify is no longer allowing new accounts to be created, and by May 1st, users won’t be able to create new stories. On May 16th, everything disappears. Storify will continue on with Storify 2, a feature of Livefyre, but will require you to purchase a Livefyre license for access. But the fact is that many users cannot or will not pay for Livefyre. Essentially, Storify will cease to exist on May 16th to most people.

So… what does this mean?

Of course, it means that you need to export anything that you have stored on Storify and want to save. (They provide instructions for exporting content on their shutting down FAQ.) More than that, however, we need to talk about how we are relying on services to archive our materials online and how that is a dangerous long-term preservation strategy.

The fact is, free Internet services can change in an instant, and without consulting their user base. As we have seen with Storify — as well as other services like Google Reader — what seems permanent can disappear quickly. When it comes to long-term digital preservation, we cannot solely depend on them as our only means of preservation.

That is not to say that we cannot use free digital tools like Storify. Storify was a great way to collect Tweets, present stories, and get information out to the public. And if you or your institution did not have the funds or support to create a long-term preservation plan, Storify was a great stop-gap until then. But digital preservation is a marathon, not a race, and we will need to continue to find new, innovative ways to ensure that digital material remains accessible.

When I heard Storify was shutting down, I went to our Scholarly Commons intern Matt Pitchford, whose research is on social media and who has a real stake in maintaining digital preservation, for his take on the issue. (You can read about Matt’s research here and here.) Here’s what Matt had to say:

Thinking about [Storify shutting down] from a preservation perspective, I think it reinforces the need to develop better archival tools along two dimensions: first, along the lines of navigating the huge amounts of data and information online (like how the Library of Congress has that huge Twitter archive, but no means to access it, and which they recently announced they will stop adding to). Just having all of Storify’s data wouldn’t make it navigable. Second, that archival tools need to be able to “get back” to older forms of data. There is no such thing as a “universally constant” medium. PDFs, twitter, Facebook posts, or word documents all may disappear over time too, despite how important they seem to our lives right now. Floppy disks, older computer games or programs, and even recently CDs, aren’t “accessible” in the way they used to be. I think the same is eventually going to be true of social media.
Matt brings up some great issues here. Storify shutting down could simply be a harbinger of more change online. Social media spaces come and go (who else remembers MySpace and LiveJournal?), and even the nature of posts change (who else remembers when Tweets were just 140 characters?). As archivists, librarians, and scholars, we will have to adopt, adapt, and think quickly in order to stay ahead of forces that are out of our control.
And most importantly, we’ll have to save backups of everything we do.

Spotlight: Unexpected Surprises in the Internet Archive

Image of data banks with the Internet Archive logo on them.

The Internet Archive.

For most of us, our introduction to the Internet Archive was the Wayback Machine, a search engine that can show you snapshots of websites from the past. It’s always fun to watch a popular webpage like Google evolve from November 1998 to July 2004 to today, but there is so much more that the Internet Archive has to offer. Today, I’m going to go through a few highlights from the Internet Archive’s treasure trove of material, just to get a glimpse at all of the things you can do and see on this amazing website.

Folksoundomy: A Library of Sound

Folksoundomy is the Internet Archives’ collection of sounds, music and speech. The collection is collaboratively created and tagged, with many participants from outside of the library sphere. There are more than 155,960 items in the Folksoundomy collection, with items that range in date back to the invention of Thomas Edison’s invention of recorded sound in 1877. From Hip Hop Mixtapes to Russian audiobooks, sermons to stand-up comedy, music, podcasts, radio shows and more, Folksoundomy is an incredible resource for scholars looking at the history of recorded sound.

TV News Archive

With over 1,572,000 clips collected since 2009, the TV News Archive includes everything from this morning’s news to curated series of fact-checked clips. Special collections within the TV News Archive include Understanding 9/11, Political Ads, and the TV NSA Clip Library. With the ability to search closed captions from US TV new shows, the Internet Archive provides a unique research opportunity for those studying modern US media.

Software Library: MS-DOS Games

Ready to die of dysentery on The Oregon Trail again? Now is your chance! The Internet Archive’s MS-DOS Games Software Library uses an EM-DOSBOX in-browser emulator that lets you go through and play games that would otherwise seem lost to time. Relive your childhood memories or start researching trends in video games throughout the years with this incredible collection of playable games!

National Security Internet Archive (NSIA)

Created in March 2015, the NSIA collects files from muckracking and national security organizations, as well as historians and activists. With over 2 million files split into 36 collections, the NSIA helps collect everything from CIA Lessons Learned from Czechoslovakia to the UFO Files, a collection of declassified UFO files from around the world. Having these files accessible and together is incredibly helpful to researchers studying the history of national security, both in the US and elsewhere in the world.

University of Illinois at Urbana-Champaign

That’s right! We’re also on the Internet Archive. The U of I adds content in several areas: Illinois history, culture and natural resources; US railroad history; rural studies and agriculture; works in translation; as well as 19th century “triple-decker” novels and emblem books. Click on the above link to see what your alma mater is contributing to the Internet Archive today!

Of course, this is nowhere near everything! With Classic TV CommercialsGrateful Dead, Community Software and more, it’s definitely worth your time to see what on the Internet Archive will help you!

This Semester at the Scholarly Commons

The sun is shining, the birds are singing, and it’s a new semester at the University of Illinois at Urbana-Champaign. And with that new semester come all of the happenings at the Scholarly Commons. We have some great things coming up!

Hours

We’re back to our normal hours. Come visit us from 9 AM – 6 PM, Monday – Friday. We hope to see you soon!

Survey Research Lab

Survey Research Lab open hours are back! Walk-ins are accepted from 2 – 5 PM every Thursday, or you can make an advance appointment by emailing Linda Owens and Karen Retzer (please copy both addresses on your email).

During Open Horus, the Survey Research Lab can look at sampling, questionnaire design, and analysis. Come in with questions about the dos and don’ts of survey wording, recommendations for designing a sampling strategy, or advice on drafting a questionnaire!

CITL Statistical Consulting

Starting January 8th and running through the end of the semester, CITL graduate students will provide free statistical consulting in the Scholarly Commons. CITL consulting will be 11 AM – 4 PM every Monday – Friday in office 306H. Consultants work with SPSS, ATLAS.ti, Stata, R, and SAS. Make an appointment for your consultation by emailing citl-data@illinois.edu.

Savvy Researcher Workshops

Our Savvy Researcher Workshop calendar is finally up! New offerings this semester include A Crash Course in Open Access and Publishing Your Research in OA, Topic Modeling Theory and Practice, Building Your Research Profile and Network, Creating Digital Books with PressBooks, Do You Know Your Fair Use Rights?, Choosing the Right Sources: Identifying Bias and Fallacies, Basics of Data Visualization, and Add Captions to Kaltura video with Automatic Speech Recognition.

Staff

The staff here at the Scholarly Commons is always ready to welcome you! Our Scholarly Commons interns, Matt Pitchford and Clay Alsup are back, as well as Megan Ozeran, our data analytics and visualization resident librarian! You can request a consultation with them or any other staff member on our Contact an Expert page.

Hope to see you soon!

Digital Timeline Tools

Everyone has a story to tell. For many of us doing work in the humanities and social sciences, presenting our research as a timeline can bring it new depth and a wider audience. Today, I’ll be talking about two unique digital storytelling options that you can use to add dimension to your research project.

Timeglider

An image of Timeglider's sample timeline on the Wright Brothers

Timeglider is an interactive timeline application. It allows you to move in and out time, letting you see time in large or small spans. It also allows events to overlap, so you can show the relationship of things in time. Timeglider also gives some great aesthetic options, including what they call their “special sauce” — the way they relate the size of an event to its importance. This option emphasizes certain events in the timeline to the user, and can make getting important ideas across simpler.

Started in 2002 as a flash-based app, Timeglader is one of the older timeline options on the web. After a major redesign in 2010, Timeglider is now written in HTML5 and JavaScript. Timeglider is free for students for a basic package, and plans for non-students can choose to pay either $5/month or $50/year.

Overall, Timeglider is an interesting timeline application with numerous options. Give it a try!

myHistro

A screenshot from a myHistro project on the Byzantine Empire.

myHistro uses text, video and pictures on maps and timelines to tell stories. Some of the power of myHistro comes from the sheer amount of information you can provide in one presentation. Presentations can include introductory text, an interactive timeline, a Google Maps-powered annotated map, and a comment section, among other attributes. The social aspect, in particular, makes myHistro powerful. You can open your work up to a large audience, or simply ask students and scholars to make comments on your work for an assignment. Another interesting aspect of myHistro is the sheer amount of projects people have come up with for it. There is everything from histories of the French Revolution to the biography of Justin Bieber, with everything in between!

myHistro is free, and you can sign up using your email or social network information.

Public Domain and Creativity

This post was guest authored by Scholarly Communication & Publishing Graduate Assistant Nicole Moriah Rhodes.


The first American copyright law protected works for fourteen years after they were published and gave the copyright owner the opportunity to renew the copyright for another fourteen years. Few did, and works passed quickly into the public domain.

The copyright term is much longer now–it varies, but you, a human, will likely own many copyrights until 70 years after you die. Some people argue that a long copyright term increases the incentive to make creative work.

However, despite the longer term, statistical analysis of the number of copyright registrations through changes in population, economy, US law, and available technology doesn’t find that increasing copyright protection increases the number of copyrighted works. Raymond Shih Ray Ku, Jiayang Sun, & Yiying Fan (2009) find that the people advocating for broader copyright laws probably aren’t advocating for an increase in the amount of creative work: the best indicator of the number of new creative works among the variables in their study is population. Their data suggest that “Laws that reduce or otherwise limit copyright protection are actually more likely to increase the number of new works” (1673) than laws granting more protection.

Such a long period of copyright protection leaves a lot of content unusable to other creators. This comic about documentary filmmakers demonstrates how stringent copyright protections can prevent creative remixing and impede the accurate representation of the world. Work in the public domain can be shared freely, but our real lives are full of content protected by copyright, and people trying to make documentaries can be inhibited by copyright even on incidental work. When they want to use copyrighted material under the fair use doctrine, the threat of lawsuits can have a chilling effect.

Lawrence Lessig (2004) uses the phrase “Walt Disney creativity” to describe “a form of expression and genius that builds upon the culture around us and makes it something different” (24). Disney’s Cinderella, Disney’s live-action Cinderella, fanfiction, and The Lizzie Bennet Diaries could all be considered examples of Walt Disney creativity. But Disney had access to fairly recent work in his time. As Lessig writes:

“Thus, most of the content from the nineteenth century was free for Disney to use and build upon in 1928. It was free for anyone— whether connected or not, whether rich or not, whether approved or not—to use and build upon.

“From 1790 until 1978, the average copyright term was never more than thirty-two years, meaning that most culture just a generation and a half old was free for anyone to build upon without the permission of anyone else. Today’s equivalent would be for creative work from the 1960s and 1970s to now be free for the next Walt Disney to build upon without permission. Yet today, the public domain is presumptive only for content from before the Great Depression.” (24-25)

Michael Hart, the creator of Project Gutenberg and a longtime Urbana resident, viewed copyright law as impeding the abundance that technology could create, beginning with the very first copyright laws after the invention of the Gutenberg Press. While Ku, Sun, & Fan (2009) do find that copyright law helps create and protect both wealth and jobs and allows creators to be rewarded for their work rather than requiring sponsorship, they advocate for reducing copyright protection where it impedes distribution or creativity.

“Because copyright law works in the negative—effectively saying ‘do not use this work, do not copy this work, do not imitate this work’—we are not sending a message that society values the creation of new works. We are only sending the message that we should stay away from those works already created” (1722).

Creative Commons is one venture designed to allow creators to share their work for other creators’ use while preserving the level of protection they choose. However, the default is still a system that restricts access to cultural works past the time when the creator might care, and can even keep works from being preserved so they will be usable when they enter the public domain. Creators should be able to benefit from the work they create, but increasing protections does not necessarily increase those benefits. Excessive copyright terms keep us from being able to discuss and rethink our common culture.

Copyright as a Tool for Censorship

This post was guest authored by Scholarly Communication & Publishing Graduate Assistant Nicole Moriah Rhodes.


Copyright should be used to encourage speech and not to silence it. The stories below demonstrate that copyright can be used to limit the rights of technology users and censor criticism.

“In practical terms, the DMCA legalized technical controls on access to electronic works; it renders obsolete traditional rules for reading and sharing print materials and, simultaneously, enables content owners to implement a pay-per-use system that controls who has access, when, how much and from where. So, for instance, you can lend a paperback to friends, but you aren’t allowed to do the same thing with an electronic book.”

“The database shows that Ares Rights has filed at least 186 complaints since 2011, with 87 made on behalf of politicians, political parties, state media, and state agencies in the Americas.” (CPJ)

“They were received by political commentators who used images of Correa, transmitted on Ecuadoran public television, in videos uploaded to YouTube, in order to make visible the resistance of local communities to the onslaught of mining communities in the country’s inland provinces. The same thing happened with videos that used stock footage to illustrate the inconsistencies of the President’s statements together with videos of protests against the exploitation of Yasuní national park, and images of repression against students.” (Derechos Digitales)

  • Electronic Frontier Foundation: To be eligible under the DMCA’s safe harbor provisions, companies must comply with legitimate takedown notices. But many hosts end up taking down content that can be legally shared. Copyright takedown notices can be used to hassle critics. Punishing bogus claims is difficult, and the damages for failing to comply can be severe.

“According to the latest numbers, Twitter does not comply with nearly 1 in 4 takedown notices it receives; Wikimedia complies with less than half; and WordPress complies with less than two-thirds. Each organization explains in its report that the notices with which they don’t comply are either incomplete or abusive.”

Closed Doors or Open Access?: Envisioning the Future of the United States Copyright Office

Copyright Librarian Sara Benson

It’s Copyright Week! For today’s theme of “transparency”, Copyright Librarian Sara Benson discusses her thoughts on the Copyright Office activities to review Section 108.


In 2005, the Copyright Office, under the guidance of the Register of Copyrights at the time, Mary Beth Peters, called for a Study Group to convene and review possible amendments to Section 108. A follow up meeting was held in 2012. These meetings were not unusual, but what followed them, was both strange and unsettling.

The procedures after the Study Group, which took place in the summer of 2016 under the guidance of Maria Pallante, were unusual in that they took place in face-to-face meetings between concerned citizens and members of the Copyright Office rather than in a call for online communications between citizens and the Office. On the one hand, this gave the members of the Office a chance to engage in a dialogue with the concerned citizens. On the other, it meant that generally only those with the resources to travel to Washington, D.C. were privileged with the ability to engage with the members of the Office. However, the Office did note that it would engage in telephone conversations, if necessary. In any event, none of these conversations were ever made public.

At that time, it seemed that the Copyright Office was making an intentional move away from a public debate about copyright to a cloistered room with a privileged few. In my view, that move was undemocratic and should be discouraged in the future. Indeed, although the Copyright Office did publish a list of individuals and organizations it met with to discuss Section 108, but the actual subject and content of those discussions remains a mystery.

Notably, shortly after taking office as the new Librarian of Congress, Dr. Carla Hayden removed Maria Pallante from her position as Register of Copyrights. Does this signal a move away from the process that was undertaken to review Section 108? Likely it does, as Librarian of Congress Dr. Hayden has recently taken further steps towards listening to the views of the multitude by openly polling the public about what we would like to see in the next Register of Copyrights.

This is an exciting time to engage with the Copyright Office under Dr. Hayden’s leadership. I encourage everyone reading this essay to add your voice to the ongoing discussions about the changes to the Office, including the selection of the new Register of Copyrights and beyond.

Got Bad Data? Check Out The Quartz Guide

Photograph of a man working on a computer.

If you’re working with data, chances are, there will be at least a few times where you encounter the “nightmare scenario”. Things go awry — values are missing, your sample is biased, there are inexplicable outliers, or the sample wasn’t as random as you thought. Some issues you can solve, other issues are less clear. But before you tear your hair out — or, before you tear all of your hair out — check out The Quartz guide to bad data. Hosted on GitHub,The Quartz guide lists out possible problems with data, and how to solve them, so that researchers have an idea of what next-steps can be when their data doesn’t work as planned.

With translations into six languages and a CreativeCommons 4.0 license, The Quartz guide divides problems into four categories: issues that your source should solve, issues that you should solve, issues a third-party expert should help you solve, and issues a programmer should help you solve. From there, the guide lists specific issues and explains how they can or cannot be solved.

One of the greatest things about The Quartz guide is the language. Rather than pontificating and making an already frustrating problem more confusing, the guide lays out options in plain terms. While you may not get everything you need for fixing your specific problem, chances are you will at least figure out how you can start moving forward after this setback.

The Quartz guide does not mince words. For example, in the “Data were entered by humans” example, it gives an example of messy data entry then says, “Even with the best tools available, data this messy can’t be saved. They are effectively meaningless… Beware human-entered data.” Even if it’s probably not what a researcher wants to hear, sometimes the hard, cold truth can lead someone to a new step in their research.

So if you’ve hit a block with your data, check out The Quartz guide. It may be the thing that will help you move forward with your data! And if you’re working with data, feel free to contact the Scholarly Commons or Research Data Service with your questions!

(Baseball) Bibliometrics: The Final Inning

Man throwing a baseball.

This post was guest authored by Scholarly Communication and Publishing Graduate Assistant Paige Kuester. This is the third part of a three-part series. Read Part 1 and Part 2.


We’re almost there! We’ve just got to go over some of the newest ways of measuring impact, and then we can all go home with that win.

Altmetrics

This is one of the newer types of bibliometrics. Instead of traditional methods of just including citations to gain an understanding of an author’s work, this “alternative metric” takes into account different ways that people interact with articles, like bookmarking them, blogging about them, tweeting, and so much more in addition to regular citations.

Different companies have different ways that they establish their scores, but here are a few that provide this service, including the company Altmetric, ImpactStory, and Plum Analytics.

Of course, there is criticism for this method also, because it does not fall strictly within the scholarly realm, and some article topics are not inherently tweetable. The scores should not be used as the sole judge of the impact of an article because there are always various factors involved. Obviously, you can also game this metric and the h-index by talking and citing your own articles, and getting your friends to do it, too. However, it is one of the new ways that these measures are trying to keep up with the technological times, as papers are more widely available online and so may have a wider audience impact than previously.

I suppose this is a bit like picking your favorite baseball player. While altmetrics are clearly less subjective than that, there still are similarities. Other bibliometrics measures only look at how other scholars use the work, which would be just like only allowing other players on your team pick your favorite player. They probably have good judgement about the ability of the player, but they are not looking at outside factors, like how the player behaves on twitter.

This points to another downside of altmetrics, that a high score only indicates a high level of attention towards an article, though this could be negative press, just like players who get a lot of news coverage when they commit crimes. Nevertheless, altmetrics does not just seem to be a trend because scholars do want to see the impact of their articles outside of strict academia, but there may still be some wrinkles to iron out.

We have covered almost all of the bases now.

Now we enter the realm of proposed metrics.

S-index

The s-index, otherwise known as the self-citation index, is exactly what is sounds like. Flatt, Blasimme, and Vayena propose this in a 2017 paper, not to replace other indices, but just to be included with the h-index so that others can get an idea as to whether an author’s h-index is self-inflated. The authors present evidence that for every self-citation, that article tends to get an average of 2-3 more citations than they would have otherwise. Arguments against this are proposed by Phil Davis from the Scholarly Kitchen. He specifically points out that not all self-citations are bad, some are necessary if the field is small and/or the author is building on their own previous research. It is not yet clear if this will catch on or not.

One way to look at this would be if baseball players only hit the ball when other players were on base and about to score. It could be a specific strategy that the player only has a successful at-bat when they have a chance to improve their RBI. But who wouldn’t take that? It’s good for them, and it is good for other players who only get to score, or get exposed to home plate because of that player. Or it could be just the situation that the player is in.

Ok, not the best example, but I promise that’s it. We made it to the end!

Not all of the metrics here are perfect. They analyze different aspects of an article or an author’s impact and so should be used in conjunction with each other. As technology opens new doors to sharing work and assessing impact, we will probably be seeing new measures and new ways to measure old metrics. Just like baseball, there is not one single number that can tell you everything you need to know, but things are improving and changing every day.

Now you know enough about the metrics of scholarly publishing to score a home run.

Or something like that.


Sources:

Allard, R. J. (2017). Measuring Your Research Impact. Uniformed Services University. Retrieved September 21, 2017, from http://usuhs.libguides.com/c.php?g=184957&p=2506307

Davis, P. (2017, September 13). Do We Need a Self-Citation Index? Scholarly Kitchen. Retrieved September 21, 2017, from https://scholarlykitchen.sspnet.org/2017/09/13/need-self-citation-index/.

De Groot, S. (2017). Measuring Your Impact: Impact Factor, Citation Analysis, and other Metrics. UIC University Library. Retrieved September 21, 2017, from http://researchguides.uic.edu/if/impact

Flatt, J. W., Blasimme, A., & Vayena, E. (2017). Improving the Measurement of Scientific Success by Reporting a Self-Citation Index. Publications 5(3). Retrieved September 21, 2017, from http://www.mdpi.com/2304-6775/5/3/20/htm

Garfield, E. (2006). The History and Meaning of the Journal Impact Factor. JAMA 295(1). Retrieved September 21, 2017, from http://garfield.library.upenn.edu/papers/jamajif2006.pdf

Reuter, A. (2017, September 11). Baseball Standing Explained. Livestrong.com. Retrieved September 21, 2017, from http://www.livestrong.com/article/337911-baseball-standings-explained/