The Social Semantic Web – a message for scholarly publishers

November 15th, 2010
by jodi

I always appreciate how Geoffrey Bilder can manage to talk about the Social Semantic Web and the early modern print in (nearly) the same breath. See for yourself in the presentation he gave to scholarly publishers at the International Society of Managing and Technical Editors last month.

Geoff’s presentation is outlined, to a large extent, in an interview Geoff gave 18 months ago (search “key messages” to find the good bits). I hope to blog further about these, because Geoff has so many good things to say, which deserve unpacking!

I especially love the timeline from slide 159, which shows that we’re just past the incunabula age of the Internet:

The Early Modern Internet

We're still in the Early Modern era of the Internet. Compare to the history of print.

Tags: , , , , , ,
Posted in future of publishing, information ecosystem, PhD diary, scholarly communication, semantic web, social semantic web, social web | Comments (3)

Accessing genomics workflows from Word documents with GenePattern

November 14th, 2010
by jodi

What if you could rerun computational experiments from within a scientific paper?

The GenePattern add-on for Word for Windows integrates reusable genomic experiment pipelines into Microsoft Word. Readers can rerun the original or modified experiments from within the document by connecting to a GenePattern server.

Rerunning a pipeline inside Word

Rerunning a pipeline inside Word

I don’t run Windows, so I took this screenshot from a video produced at the Broad Institute of MIT and Harvard, where GenePattern is developed.

Readers without Word for Windows can also access the experimental pipelines by exporting them from the document: just run a GenePatternDocumentExtractor command from a GenePattern server. The GenePattern public server was very easy to access and start using. Here’s what the GenePatternDocumentExtractor command looks like:

Running GenePatternDocumentExtractor at the GenePattern public server

Running GenePatternDocumentExtractor at the GenePattern public server

Unfortunately the jobs I ran didn’t extract any pipelines from the Institute’s sample DOC. I’ve sent in an inquiry (either I’m doing something wrong or there’s a bug, either way it’s useful). I was very impressed that I could make my jobs public, then refer to them by URL in my email, to make clear what exactly I did.

The GenePattern add-on for Word is another find from the beyondthepdf list. Its development was funded by Microsoft. See also Accessible Reproducible Research by Jill P. Mesirov (Science, 327:415, 2010). doi:10.1126/science.1179653, which describes the underlying philosophy: have a Reproducible Research System (RRS) made up of an environment for doing computational work (the Reproducible Research Environment or RRE) and an authoring environment (the Reproducible Research Publisher or RRP) which links back to the research system.

Tags: , , , , , ,
Posted in books and reading, future of publishing, information ecosystem, scholarly communication | Comments (1)

Utopia Documents: pulling scientific data into the PDF for interactive exploration

November 14th, 2010
by jodi

What if data were accessible within the document itself?

Utopia Documents is a free PDF viewer which recognizes certain enhanced figures, and fetches the underlying data. This allows readers to view and experiment with the tables, graphs, molecular structures, and sequences in situ.


You can download Utopia Documents for Mac and Windows to view enhanced papers, such as those published in The Semantic Biochemical Journal.

These screencasts were made from pages 9 and 10 of PDF of a paper by the Manchester-based Utopia team: T. K. Attwood, D. B. Kell, P. Mcdermott, J. Marsh, S. R. Pettifer, and D. Thorne. Calling international rescue: knowledge lost in literature and data landslide! Biochemical Journal, Dec 2009. doi:10.1042/BJ20091474.

In an interview at the Guardian, Utopia’s Phillip McDermott says:

“Utopia Documents links scientific research papers to the data and to the community. It enables publishers to enhance their publications with additional material, interactive graphs and models. It allow the reader to access a wealth of data resources directly from the paper they are viewing, makes private notes and start public conversations. It does all this on normal PDFs, and never alters the original file. We are targeting the PDF, since they still have around 80% readership over online viewing.

“Semantics, loose-coupling, fingerprinting and linked-data are the key ingredients. All the data is described using ontologies, and a plug-in system allows third parties to integrate their database or tool within a few lines of script. We use fingerprinting to allow us to recognise what paper a user is reading, and to spot duplicates. All annotations are held remotely, so that wherever you view a paper, the result is the same.”

I’d still like to see a demo of the commenting functionality.

I’d also be particularly interested in the publisher perspective, about the production work that goes into creating the enhancements. Portland Press’s October news announces that they’ve been promoting Utopia at the Charleston conference and SSP, with an upcoming appearance at the STM Innovations Seminar.

Utopia came to my attention via Steve Pettifer’s mention.

Tags: , , , , , , , , ,
Posted in future of publishing, information ecosystem, library and information science, scholarly communication, semantic web, social semantic web | Comments (4)

A Model-View-Controller perspective of scholarly articles

November 13th, 2010
by jodi

A scholarly paper is not a PDF. A PDF is merely one view of a scholarly paper. To push ‘beyond the PDF’, we need design patterns that allow us to segregate the user interface of the paper (whether it is displayed as an aggregation of triples, a list of assertions, a PDF, an ePub, HTML, …) from the thing itself.

Towards this end, Steve Pettifer has a Model-View-Controller perspective on scholarly articles, which he shared in a post on the Beyond the PDF listserv, where discussions are leading up to a workshop in January. I am awe-struck: I wish I’d thought of this way of separating the structure and explaining it.

I think a lot of the disagreement about the role of the PDF can be put down to trying to overload its function: to try to imbue it with the qualities of both ‘model’ and ‘view’. … One of the things that software architects (and I suspect designers in general) have learned over the years is that if you try to give something functions that it shouldn’t have, you end up with a mess; if you can separate out the concerns, you get a much more elegant and robust solution.

My personal take on this is that we should keep these things very separate, and that if we do this, then many of the problems we’ve been discussing become more clearly defined (and I hope, many of the apparent contradictions, resolved).

So… a PDF (or come to that, an e-book version or a html page) is merely a *view* of an article. The article itself (the ‘model’) is a completely different (and perhaps more abstract) thing. Views can be tailored for a particular purpose, whether that’s for machine processing, human reading, human browsing, etc etc.

[paragraph break inserted]

The relationship between the views and their underlying model is managed by the concept of a ‘controller’. For example, if we represent an article’s model in XML or RDF (its text, illustrations, association nanopublications, annotations and whatever else we like), then that model can be transformed in to any number of views. In the case of converting XML into human-readable XHTML, there are many stable and mature technologies (XSLT etc). In the case of doing the same with PDF, the traditional controller is something that generates PDFs.

[paragraph break inserted]

The thing that’s been (somewhat) lacking so far is the two-way communication between view and model (via controller) that’s necessary to prevent the views from ossifying and becoming out of date (i.e. there’s no easy way to see that comments have been added to the HTML version of an article’s view if you happen to be reading the PDF version, so the view here can rapidly diverge from its underlying model).

[paragraph break inserted, link added]

Our Utopia software is an attempt to provide this two-way controller for PDFs. I believe that once you have this bidirectional relationship between view and model, then the actual detailed affordances of the individual views (i.e. what can a PDF do well / badly, what can HTML do well / badly) become less important. They are all merely means to channeling the content of an article to its destination (whether that’s human or machine).

The good thing about having this ‘model view controller’ take on the problem is that only the model needs to be pinned down completely …

Perhaps separating out our concerns in this way — that is, treating the PDF as one possible representation of an article — might help focus our criticisms of the current state of affairs? I fear at the moment we are conflating the issues to some degree.

– Steve Pettifer in a Beyond the PDF listserv post

I’m particularly interested in hearing if this perspective, using the MVC model, makes sense to others.

Tags: , , , , , , ,
Posted in books and reading, future of publishing, information ecosystem, library and information science, scholarly communication, social semantic web | Comments (9)

A Taxonomy for Decisions

November 4th, 2010
by jodi

Tim van Gelder provides a taxonomy for decisions:

  1. Intuitive Decisions
  2. Technical Decisions
  3. Deliberative Decisions
  4. Bureaucratic Decisions

Deliberative and bureaucratic decisions are, I think, the most important for collaborative decision-making. Intuitive decisions, made quickly by an individual, are least important for collaboration. Technical decisions have the most interesting description: they are “made by following some well-defined technical procedure”; arguably they are not decisions.

Can you spot any overlaps or gaps? Discuss at his article.

The argumentation community has given a lot of attention to deliberation; I wonder if that has been influenced by the prevalence of deliberation in decision-making, and the difficulty of formal modelling of bureaucracies.

Tags: ,
Posted in argumentative discussions, PhD diary | Comments (0)

blog ‘reactions’

November 2nd, 2010
by jodi

Instead of enabling commenting on your blog, you can let readers ‘react’ by marking the post as ‘funny’, ‘interesting’, or ‘cool’. So far I’ve only seen this on one Blogspot blog, Galway Library’s blog.

Reactions to a blog post

Is this post funny, interesting, or cool?


If you know whether there’s a plugin doing this, or if it’s a general (optional) Blogspot feature, please let me know in the comments.

Tags: , , ,
Posted in argumentative discussions, PhD diary, social web | Comments (1)

Ebook pricing fail (Springer edition)

November 1st, 2010
by jodi

I wrote Springer to ask about buying an ebook that’s not in our university subscriptions. They sell the print copy at €62.95, but the electronic copy comes to €425, chapter by chapter.

Publishers: this is short-sighted (not to mention frustrating)–especially when your customers are looking for a portable copy of a book they already owns!

———- Forwarded message ———-
From: Springerlink, Support, Springer DE
Date: Fri, Oct 29, 2010 at 8:46 PM
Subject: WG: ebook pricing

Dear Jodi,

Thank you for your message.

On SpringerLink you can purchase online single journal articles and book chapters, but no complete ebooks.
eBooks are sold by Springer in topical eBook packages only.

with kind regards,
SpringerLink Support Team
eProduct Management & Innovation | SpringerLink Operations
support.springerlink@springer.com | + 49 (06221) 4878 743
www.springerlink.com

—–Original Message—–
From: Jodi Schneider
Sent: Thursday, October 28, 2010 5:09 PM
To: MetaPress Support
Subject: ebook pricing

Hi,

I’m interested in buying a copy of [redacted] as an ebook:
http://www.springerlink.com/content/[redacted]

This book has 17 chapters, which seem to be priced at 25 EUR each = 425 EUR.

But I could buy a print version, new at springer.com for 62.95 EUR:
http://www.springer.com/mathematics/book/[redacted]

Can you help me get the ebook at this price?
Thanks!
-Jodi

Tags: , , , ,
Posted in books and reading, future of publishing | Comments (3)

“Like” and its misuse

October 20th, 2010
by jodi

Language evolves, and we use words loosely. But I’m more and more disturbed with the way “Like” is being manhandled.
A misuse of the Like button
Argumentation will need to encompass polarity; so I hope that it can help.

Tags: , , , ,
Posted in argumentative discussions, PhD diary, random thoughts | Comments (4)

CiTO in the wild

October 18th, 2010
by jodi

CiTO has escaped the lab and can now be used either directly in the CiteULike interface or with CiteULike machine tags. Go Citation Typing Ontology!

In the CiteULike Interface

To add a CiTO relationship between articles using the CiteULike interface, both articles must be in your own library. You’ll see a a “Citations (CiTO)” section after your tags. Click on edit and set the current article as the target.

set the CiTO target

First set the CiTO target

Then navigate around your own library to find a related article. Now you can add a CiTO tag.

Adding a CiTO tag in CiteULike

Adding a CiTO tag in CiteULike

There are a lot of choices. Choose just one. :)

CiTO Object Properties appear in the dropdown

CiTO Object Properties now appear in the dropdown

Congratulations, you’ve added a CiTO relationship! Now mousing over the CiTO section will show details on the related article.

CiTO result

Mouse over the resulting CiTO tag to get details of the related article

Machine Tags

Machine tags take fewer clicks but a little more know-how. They can be added just like any other tag, as long as you know the secret formula: cito--(insert a CiTO Object Property here from this list)--(insert article permalink numbers here) Here are two more concrete examples.

First, we can keep a list of articles citing a paper. For example, tagging an article

cito--cites--1375511

says “this article CiTO:cites article 137511”. Article 137511 can be found at http://www.citeulike.org/article/137511, aka JChemPaint – Using the Collaborative Forces of the Internet to Develop a Free Editor for 2D Chemical Structures. Then we can get the list of (hand-tagged) citations to the article. Look—a community generated reverse citation index!

Second, we can indicate specific relationships between articles, whether or not they cite each other. For example, tagging an article

cito--usesmethodin--423382

says “this item CiTO:usesmethodin item 42338”. Item 42338 is found at http://www.citeulike.org/article/423382, aka The Chemistry Development Kit (CDK):  An Open-Source Java Library for Chemo- and Bioinformatics.

Upshot

Automation and improved annotation interfaces will make CiTO more useful. CiTO:cites and CiTO:isCitedBy could used to mark up existing relationships in digital libraries such as ACM Digital Library and CiteSeer, and could enhance collections like Google Books and Mendeley, to make human navigation and automated use easier. To capture more sophisticated relationships, David Shotton has hopes of authors marking up citations before submitting papers; if it’s required, anything is possible. Data curators and article commentators may observe contradictions between papers, or methodology reuses; in these cases CiTO could be layered with an annotation ontology such as AO in order to make the provenance of such assertions clear.

CiTO could put pressure on existing publishers and information providers to improve their data services, perform more data cleanup, or to exposing bibliographies in open formats. Improved tools will be needed, as well as communities that are willing to add data by hand, and algorithms for inferring deep citation relationships.

One remaining challenge is aggregation of CiTO relationships between bibliographic data providers; article identifiers such as DOI are unfortunately not universal, and the bibliographic environment is messy, with many types of items, from books to theses to white papers to articles to reports. CiTO and related ontologies will help explicitly show the bibliographic web and relationships between these items, on the web of (meta)data.

Further Details

CiTO is part of an ecosystem of citations called Semantic Publishing and Referencing Ontologies (SPAR); see also the JISC Open Citation Project which is taking bibliographic data to the Web, and the JISC Open Bibliography Project. For those familiar with Shotton’s earlier writing on CiTO, note that SPAR breaks out some parts of the earlier formulation of this ontology.

Tags: , , , , ,
Posted in argumentative discussions, books and reading, information ecosystem, library and information science, PhD diary, scholarly communication, semantic web | Comments (3)

Enabling a Social Semantic Web for Argumentation (defining my Ph.D. research problem)

July 23rd, 2010
by jodi

I’m working on online argumentation: Making it easier to have discussions, get to consensus, and understand disagreements across websites.

Here are the 3 key questions and the most closely related work that I’ve identified in the first 9 months of my Ph.D.

Read on, if you want to know more. Then let me know what you think! Suggestions will be especially helpful since I’m writing my first year Ph.D. report, which will set the direction for my second year at DERI.


Enabling a Social Semantic Web for Argumentation

Argumentative discussions occur informally throughout the Web, however there is currently no way of bringing together all of the discussions on a given topic along with an indication of who is agreeing and who is disagreeing. Thus substantial human analysis is required to integrate opinions and expertise to, for instance, determine the best policies and procedures to mitigate global warming, or the recommended treatment for a given disease. New techniques for gathering and organising the Social Web using ontologies such as FOAF and SIOC show promise for creating a Social Semantic Web for argumentation.

I am currently investigating three main research questions to establish the Social Semantic Web for argumentation:

  1. How can we best define argumentation for the Social Semantic Web, to isolate the essential problems? We wish to enable reasoning with inconsistent knowledge, to integrate disparate knowledge, and identify consensus and disputes.  Similar questions and techniques come up in related but distinct areas, such as sentiment analysis, dialogue mapping, dispute resolution, question-answering and e-government participation.
  2. What sort of modular framework for argumentation can support distributed, emergent argumentation — a World Wide Argumentation Web? Some Web 2.0 tools, such as Debatepedia, LivingVote, and Debategraph, provide integrated environments for explicit argumentation. But our goal is for individuals to be able to use their own preferred tools — in a social environment — while understanding what else is being discussed.
  3. How can we manage the tension between informality and ease of expression on the one hand and formal semantics and retrievability/reusability on the other hand? Minimal integration of informal arguments requires two pieces of information: a statement of the issue or proposition, and an indication of polarity (agreement or disagreement). How can we gather this information without adding cognitive overhead for users?

Related Work

Ennals et al. ask: ‘What is disputed on the Web? (Ennals 2010b). They use annotation and NLP techniques to develop a prototype system for highlighting disputed claims in Web documents (Ennals 2010a). Cabanac et al. find that two algorithms for identifying the level of controversy about an issue were up to 84% accurate (compared to human perception), on a corpus of 13 arguments. These are useful prototypes of what could be done; Ennals prototype is indeed a Web-scale system, but disputed claims are not arguments.

Rahwan et al. (2007) present a pilot Semantic Web-based system, ArgDF, in which users can create arguments, and query to find networks of arguments. ArgDF is backed with the AIF-RDF ontology, and uses Semantic Web standards.  Rahwan (2008) surveys current Web2.0 tools, pointing out that integration between these tools is lacking, and that only very shallow argument structures are supported; ArgDF and AIF-RDF are explained as an improvement. What is lacking is uptake in end-user orientated (e.g. Web 2.0) tools.

The Web2.0 aspect of the problem is explored in several papers, including Buckingham Shum (2008), which presents Cohere, a Web2.0-style argumentation system supporting existing (non-Semantic Web) argumentation standards, and Groza et al. (2009) which proposes a abstract framework for modeling argumentation. These are either minimally implemented frameworks or stand-alone systems which do not yet support the distributed, emergent argumentation envisioned, as further elucidated by Buckingham Shum (2010).

References with links to preprints

  1. S. Buckingham Shum, “Cohere: Towards Web 2.0 Argumentation,” Computational Models of Argument – Proceedings of COMMA 2008, IOS Press, 2008.
  2. S. Buckingham Shum, AIF Use Case: Iraq Debate, Glenshee, Scotland, UK: 2010. http://projects.kmi.open.ac.uk/hyperdiscourse/docs/AIF-UseCase-v2.pdf
  3. G. Cabanac, M. Chevalier, C. Chrisment, and C. Julien, “Social validation of collective annotations: Definition and experiment,” Journal of the American Society for Information Science and Technology, vol. 61, 2010, pp. 271-287.
  4. R. Ennals, B. Trushkowsky, and J.M. Agosta, “Highlighting Disputed Claims on the Web,” WICOW 2010, Raleigh, North Carolina: 2010.
  5. R. Ennals, D. Byler, J.M. Agosta, and Barboara Rosario, “What is Disputed on the Web?,” WWW 2010, Raleigh, North Carolina: 2010.
  6. T. Groza, S. Handschuh, J.G. Breslin, and S. Decker, “An Abstract Framework for Modeling Argumentation in Virtual Communities,” International Journal of Virtual Communities and Social Networking, vol. 1, Sep. 2009, pp. 35-47. 
  7. I. Rahwan, “Mass argumentation and the semantic web,” Web Semantics: Science, Services and Agents on the World Wide Web, vol. 6, Feb. 2008, pp. 29-37.
  8. I. Rahwan, F. Zablith, and C. Reed, “Laying the foundations for a World Wide Argument Web,” Artificial Intelligence, vol. 171, Jul. 2007, pp. 897-921.

Tags: ,
Posted in argumentative discussions, PhD diary, semantic web, social semantic web, social web | Comments (1)