Live-blogging the 2009 Vancouver PKP Conference

Category — Technical Themed Sessions

Copenhagen Business School Library offers OCS as Conference Management System to Their Faculty: the Session Blog

Presenters: Kirsten Suhr Jacobsen, Helle Damgaard Andersen, Peder Lærke Nielsen

July 9th, 2009 at 3:00 pm

Session Abstract

Background

Helle Andersen has managed the Copenhagen Business School‘s webservices since 1999, including the main CBS page and related services such as the open conference system (OCS).  Kirsten Jacobsen has managed several international conferences hosted by CBS.  Together with Peder Nielsen, the library IT department leader, CBS offers conference management system OCS as part of the schools public web system.

Session Overview

The CBS has approximately 15,000 students and 1,000 staff making it one of the largest institutions in Northern Europe.  The school understands that conferences are a possible forum to attract investment, scholarship and students.  In turn, a professional conference system, may attract an increasing number of conferences to Denmark.

Copenhagen Business School

Copenhagen Business School

To help this initiative, the university has allocated 3 years of funding for a conference Secretariat at CBS.  A large number of conference management systems (CMS) were considered, and OCS was eventually chosen for being open-source, flexible, exclusively web-based  and with a large available community to provide technical support.

OCS is a comprehensive package that addresses the many facets of conference management, from online submissions of relevant content to management of all people involved (from registrants to presenters and organizers).  OCS even facilitates development of conference webservices and online discussion forums.  In fact, the PKP Scholarly Publishing Conference 2009, was managed using OCS.  The system’s main strengthes lies in its flexibility and  detailed indexing capactiy, which allows for dealing with multiple conferences simultaneously.  OCS 2.0 is originally based on open-source code developed for open journal systems, and the CBS group made several productive improvements through formatting modifications and software plug-ins.

Open Conference Systems Schematic

Open Conference Systems Schematic

Unfortunately, the CBS group has run into multiple difficulties overcoming template configuration problems (especially with navigation menus), and a large portion of time was spent discussing multiple suggestions for improvement to the core software.  Majority of the CBS group’s concerns described issues surrounding formatting, style design, and the lack of descriptors from pull-down menus.  Other issues raised included expansion of administrative modules, and further development of OCS documentation and tutorials.  CBS has already started this process, and suggested that perhaps resources be pooled with the PKP support group and collaborative documents be published on wikipedia.

Peder Neilsen facetiously commented that it is a “Danish tradition of not talking about what is good, but what could be better.” Certainly the described OCS strengths lay not only in its cost (free), but also in the flexibility conferred by open-source code.  Furthermore, OCS users may benefit from accessing not only the development team, but a support group comprising of an international community of fellow users.  Certainly, this session highlighted the productive interplay between the open-source software developers (of whom Alec Smecher was present in the audience to represent) and the users.  Continual feedback drives steady modifications to the original source-code, new plug-in creations and  essentially product evolution.


July 13, 2009   Comments Off on Copenhagen Business School Library offers OCS as Conference Management System to Their Faculty: the Session Blog

Importing Backissues into OJS: Development of an OJS Import Script with Django: the Session Blog

Friday, July 10th, 2009 (11:00 AM)
SFU Harbour Centre (Sauder Industries Room 2270)

Presenter: Syd Weidman, Library Systems Supervisor, University of Winnipeg – Session Abstract

Session Overview

Why was this an issue?

With the transition to open-access publishing of several journals at the University of Winnipeg, Syd Weidman and the University library have been involved in multiple aspects of this transformation.  Given that these journals have been in print for decades, one of the major obstacles that needed to be addressed was the importation of back issues into an online, open-access compatible format.

Initial attempts using the available software proved difficult.  They were met with bugs and their associated patches; overall, Syd described the process as “laborious and convoluted”.  He surmised that “in the context of importing [a large volume of] back issues, small efficiencies [may] have a large impact.”  With this notion in mind, Syd began work on the Open Journal Systems (OJS) Import Project.

Tackling the problem – Use of Django

Syd highlighted the basic design goals of any software to be used for this purpose; he stressed that the process needed to be as EASY as possible.  He sought to optimize the software’s ease of CONSTRUCTION, USE, DEPLOYMENT and MAINTENANCE.  Being most familiar and comfortable with the Python programming language, Syd opted to use the Django Web framework to build a Web-based application to carry out the task of importing back issues.

Django is an open-source framework that was initially used by the online publishing industry.  With a short digression, Syd took a moment to review the “4 freedoms” of open-source software, being the “free” use of software for any purpose, free access to its code, freedom to modify as well as an understanding that improvements will be shared with others (for more, take a look at the Free Software Foundation’s website.)  Django, in particular, has several advantages over other similar frameworks, namely:

  • object relational mapping – allows use of fewer lines of programming, increasing robustness
  • automatic administrator interface
  • elegant URL design
  • pluggable template system
  • flexible and robust cache system
  • i18n compatible – allows for the application to be adapted to other languages without significant engineering changes
  • excellent documentation
  • an active mailing list (a double-edged sword!)

Success!

With the development of the new importation software, the U of W was able to scan backissues into .pdf format, to ultimately be uploaded into their respective online journals.  This required the entering of appropriate metadata in order to allow for accurate archiving and searching.

Challenges and future directions

One of the difficulties in developing a script for another piece of software is to ensure that they remain in sync when new versions appear.  In a subsequent OJS release following the development of the OJS importing application, incompatibilities/bugs appeared, and needed patching.

Commentary/Questions

Just prior to the question period, Syd mentioned the recent development of another application, “Quick Submit”, which may now be able to perform similar functions to his program.

Related Links

University of Winnipeg library (and their OA publications: Canadian Bulletin of Medical History, Journal of Mennonite Studies and the Canadian Children’s Literature Journal)
Python programming language
Django framework

References

Weidman, S. (2009). Importing backissues into ojs: development of an ojs import script with django. PKP Scholarly Publishing Conference 2009. Retrieved 2009-07-08, from http://pkp.sfu.ca/ocs/pkp/index.php/pkp2009/pkp2009/paper/view/190

July 11, 2009   Comments Off on Importing Backissues into OJS: Development of an OJS Import Script with Django: the Session Blog

Maximizing the Reach of a Graduate Research Journal Through Open Source Tools and Social Media: The Session Blog

Presenter: Rajendra Bose

July 10th, 2009 at 3:30 p.m.

(used with permission from RB)

(used with permission from Raj Bose)

Background Info

Read the Abstract

Cultural Formations is an online journal (due to launch next month) that is open-access and peer-reviewed.  The Center for Digital Research and Scholarship (CDRS) collaborated with Columbia University’s Teachers College to form Students for a Cultural Studies Initiative (SCSI).  This project has grown to become more than just an online journal, using the concepts of Web 2.0 and Social Media as well as a research repository, Academic Commons.

As quoted on the Cultural Formations website, the journal-plus project has “…three integrated elements: an online journal of peer-reviewed articles, a blog for discussion of published articles and works in progress, and a wiki to serve as a “Wikipedia” of cultural studies and education.”

Session Overview

Graduate students at Columbia University and the Teachers College joined forces to form SCSI.  Over the past year, they have worked hard to start a “journal-plus” –  an online space that looks at cultural formations including youth culture, media, music, art, film and ethnicity, etc.   Their goal is to not only publish research but to be accessible to the entire community, not just academic scholars.  Along the way they were faced with significant challenges (examples include web design, technical problems with layout of the vastly different programs and challenges from having multiple partners with decision-making authority).   They are however on schedule to launch their site next month. (Please stay tuned to this blog for the launch announcement!)

Services they hope to provide

  • Publications – e.g. online books,  journals
  • Conferences
  • Academic Commons
  • Wikischolars

An example of what their website will look like:

(Used with permission from Raj Bose)

(Used with permission from Raj Bose)

The goals of the Cultural Formations site are three-fold:

  • authority of a journal
  • immediacy of a blog
  • what’s new and emerging in wiki format
(Used with permission from Raj Bose)

(Used with permission from Raj Bose)

The site is scheduled to launch in August 2009 at Culturalformations.columbia.edu (URL may change prior to launch).

Questions from the audience

1. Impressive. Will you open this up to the public as open source?
Yes, we are using all open access materials. The idea is to share it with the rest of the community, Users may have to connect with us to get a user profile as it may be too expensive to allow it open to all users at once. There likely will be a basic free level, and we may charge for some of the more advanced functions. We feel the site will have longevity with the article content/repository.

2. Is there a similar movement in other universities?
Yes, but the idea right now is to work with university presses and authors right here to get the right agreements.

3. Is there going to be any co-ordination with other groups?
Already there are 50-100 student journals with OJS [Open Journal System]. You especially want this system for longevity when a student group starts to fade out, this will help. The university hosting this site will help keep those groups up and running.

4. What about groups with other levels of education and how this may be used?
This idea is not a typical journal, but the idea to connect a blog and a wiki so that other members of community can join or help is amazing and very inclusive. Because of this, there will only be a light peer review. They partners want to keep this as open as possible.

Related Links

Publication List

Center for Digital Research and Scholarship (CDRS)

YouTube Preview Image

(Source)

Open Archives Initiative Object Reuse and Exchange (OAI-ORE)

July 10, 2009   Comments Off on Maximizing the Reach of a Graduate Research Journal Through Open Source Tools and Social Media: The Session Blog

Enriching Digital Citation Networks using Web 2.0 Principles: The Session Blog

Presenter: James Williams, Session Abstract

July 10, 2009 at 4:30 p.m.

james-williams-small

Photo taken at PKP 2009, with permission

Background
James Williams is Manager, Information Systems at the eScholarship Research Centre at The University of Melbourne (http://www.esrc.unimelb.edu.au). His background is in Computer Science and Linguistics.  James’ work focuses on eResearch and eScholarship, archiving and print on demand.

Session Overview

This session was a technical session, which laid the context and groundwork for a plug-in aimed at leveraging the network for citations. This session also highlights that some of the features of this plug-in are about to be implemented in the Referral plug-in coming with OJS 2.3. The author is planning to take his plug-in further by building on the work that has been done with the OJS Referral plug-in.

The context and groundwork included: Scholarly Contribution and Statistical Measurement; Bibliometrics – the h-Index and Journal Impact Factor; Indexes, Citators, Aggregators, etc.; What is Scholarship? What is Significant?; Developing Online Research Support Infrastructure; Scope – Scholarly Publishing and OJS Development. The key driver behind the paper had been to lead up to a plug-in which could capture and expose citation data, though this has been, to a certain extent, scooped by the Referral plug-in in the coming OJS 2.3. So OJS created a “heart broken Australian”, though one who is expected to make a full recovery.

In Australia, extensive dialogue around the measurement of academics’ scholarly research. Much of this is controversial. One key measure discussed is citation measures. Two citation measures come up in these discussions: the h-index and the journal impact factor. Many questions have been raised about this and attempts made to improve these measures, but questions remain about the extent to which they can be relied upon. The h index is from a paper by Hirsh, (2005). Crudely, it is the average number of citations that an author’s publications achieved in a year. Though its calculation is more complex than this.

Further problems with citation measures include statistical considerations with these measures which vary greatly between disciplines, so citations counts can vary greatly for statistical reasons. The source of these citations is the different indexes (Scopus, ISI Web Of Science, PubMed, Google Scholar) though each of these measures citations differently.

So, in general, what is needed is to develop intelligent digital citation networks, increase the discoverability of significant research contributions, harness the footprints caused by active network behaviour, provide greater statistical detail for transparency, and surface this information as much as possible for richer bibliometric reporting.

So, more specifically, what is needed is to track referrer information where it’s available: referring  website/link/journal, date of referral and number of incoming hits from each referrer. As each referring URL,  site, or journal is detected, it then becomes a known information source – an external, digital citation. We then surface this URL, site, or journal on the originating journal. Users of our (originating) journal, unaware of the external citation, can then find other people referring to this resource, and so, the networked community expands, develops, and evolves. This is the intent behind our planned digital citation plug-in, but which, as previously mentioned has been partially scooped by the Referral plug-in in OJS 2.3. So we are going to look at the Referral plug-in and talk further with PKP about this and build on the work that PKP has done.

Session Questions

Q. I’d like to know more about the h index?
A: Hirsh (2005) and others have expanded this. Peter Jascoe has also written a paper on this

Q. What software should use would where citations need to be checked in the publication process? How can we auto check and fix citations?
A. For OJS, the Resolver plug-in will help. L8X could also help.

References and Related Links

eScholarship Research Centre, University of Melbourne

Hirsh, J.E. (2005). An index to quantify an individual’s scientific research output. Retrieved on July 12, 2009 from http://www.cs.ucla.edu/~palsberg/hirsch05.pdf

OJS Plugins

Top Scholars in Computer Science, as calculated by the h index

What is Web 2.0? (includes translations in Chinese, French, German, Italian, Japanese, Korean, Spanish)

July 10, 2009   Comments Off on Enriching Digital Citation Networks using Web 2.0 Principles: The Session Blog

The Latest Developments in XML Content Workflows: The Session Blog

Presenter: Adrian Stanley, Session Abstract

July 10, 2009 at 11:00 a.m.

Background

Adrian Stanley is the Chief Executive Officer for The Charlesworth Group (USA). Prior to this Adrian worked for 4 year as Production Director for Charlesworth China setting up their Beijing office. The Charlesworth Group offers cutting edge automated typesetting services, as well as Rights and Licensing opportunities for publishers in the China market. Adrian is an active committee member for the SSP (Society for Scholarly Publishers), The Association of Learned and Professional Society Publishers (ALPSP North American Chapter), the Council of Science Editors (CSE), as well as working on a key project with the Canadian Association of Learned Journals; he has 20 years experience working in the publishing/printing sector.

Session Overview

This session was a technical session, yet one which tried to keep the technology basic and focused more on the benefits of the technology for the publishing process, particularly the benefits of XML. Charlesworth works with XML and XML workflows. Their aim is to support publishers. This session showcased their tools.

The benefits of using XML are the automation and time saving in the publisher’s workflows. But more fundamentally XML adds energy to data. The presenter took a short poem and showed it as text, then HTML then XML. The change in the poem was less how it looked than what it now was. In XML, it now had substantial additional information encoded with the poem.

Benefits of XML are numerous, adding more information and data about an article, data that is both machine and human readable. The presenter referred to the excellent pre-conference workshop in the PKP 2009 conference by Juan Pablo Alperin for further information on XML.

autoproof_process

Image taken from presenter's presentation at PKP 2009, with permission

AutoProof Process Timeline diagram highlights the 10 minute process that Charlesworth’s AutoProof performs. It takes a styled file, converts it to XML and then outputs it in production ready PDF format. Charlesworth provides some Microsoft Word macros that help apply tagging and styling, to prepare documents in Microsoft Word for import to this process and ultimately XML. The presenter took us through this process showing each stage.

The presenter then showed the Online Tracking System that allows a view into each stage in the AutoProof process. From one screen, the publisher and editor can run the entire publishing process.

Next some examples of the many other types of AutoProof publications that are possible were shown. It’s not just journals e.g. program and abstract meetings, ebooks, dictionaries, etc can be handled. As long as files coming to AutoProof are structured, many types of documents can be used and created. So what AutoProof is doing is simplification and reduction of time in the workflow process, using XML at its core.

In addition AutoProof supports incorporating XML into the PDF, to make PDFs more readable and searchable. Storing the metadata in the PDF file (using XMP) could make it much easier to import a large collection of PDF files into a reference manager. XMP combines XML metadata within the PDF file. The presenter showed a sample XML Packet (XMP) page.

Other applications and developments with XML are also possible including auto generation of table of contents and Index, author proofing of link to PDF form XML, creating XML in multiple DTDs (document type definition), etc.

So in summary the goals of AutoProof are to enhance publications, customize how much or how little the publisher wants to do with XML, provide integrations with other systems and provide the basis for further developments e.g. authoring templates. Lastly, the presenter showed an example of OJS journal using AutoProof. Every reference automatically picked up the DOI (Digital Object Identifier) and automatically creates web links to the references.

Session Questions
Q: How much of publishing workflow did the James Journal use?
A: The James Journal allowed forward and backward linking, see the references in this James Journal article. They used most of our services; most fundamentally we created the XML. Because they used a lot of our services they were able to publish this magazine with a part-time editor.

Q: What about integration with OJS?
A: So we are working on creating integration with OJS

Q: What are the costs?
A: Broad range, depending on services that you need. The important point however, is to start to work out how much the alternatives of using your staff time is in fact really costing you. That is where we can help you make savings by significantly reducing the amount of staff time you need to dedicate to the publication process.

References and Related Links

AutoProof

AutoProof Overview Leaflet

Charlesworth Group

Charlesworth Group launches new open access OJS journal

DTDs (document type definition)

DOI (Digital Object Identifier)

The James Journal

XML-Based Workflow and Process Management Standards

XMP

July 10, 2009   Comments Off on The Latest Developments in XML Content Workflows: The Session Blog

Socializing and Disseminating the Academic and Intellectual Creation: Experiences from La Plata National University: The Session Blog

Date: July 10, 2009

Presenters: Gonzalo Villarreal, Marisa Raquel De Giusti

s13610, “Feet” January 27, 2009 via Flickr, Creative Commons Attribution.

s13610, “Feet” January 27, 2009 via Flickr, Creative Commons Attribution. La Plata has a sandbox for OJS and OCS users to get a "feel" with templates!

Session Overview

Abstract

This presentation was delivered by Gonzalo in real time through Skype.

Background on La Plata National University (UNLP)

He described UNLP as a major university in Argentina and Latin America. It has more then 90 000 students and 10 000 employees, including staff and administration.

About UNLP:

They produce journal articles, degree essays and theses, multimedia production, books and e-books and other products.

History of Development

In 2003, UNLP launched the Intellectual Creation Dissemination Service, SeDiCi, for 2 purposes:
– to provide authors a means to publish
– to benefit society for all it has invested in the university

The original plan of the university was to receive, digitize, catalog and upload documents to the SeDiCi website, but software was needed. After an extensive worldwide search, the team felt they had to design their own software. Celsius-DL was the answer to complete the service to scholars. It has since been recognized as the OEA’s award in 2005 as the best Digital Library in America.

Celsius-DL supports a free scheme and is used to configure a wide range of  materials. It exports to Dublin-Core or Marc21, is a data and service provider and it is built on open source technologies.

Since 2007, UNLP has recognized the importance of disseminating academic and intellectual creations. Journals also were seen as crucial to disseminate, so SeDiCi was expanded to include them. In October, 2008, UNLP launched the Journals Portal. Pre-existing and new publications are both accepted. Authors publish through a review process, easily communicating with other authors, reviewers and editors. OJS was chosen as the platform for them to use.

They then decided to create the Congresses Portal. This helps groups and others organize meetings. PKP’s OCS was selected to supply the service. SeDiCi’s team offers user, systems and technical support to users, but they had to overcome some obstacles and learn how to use it.

Obstacles and Solutions

  1. Obstacles included many words or phrases in Argentina that were different from Spain. Corrected translations are now complete. In service to others, they are also available on PKP’s website for download.
  2. In addition, some OCS users needed to use some LaTex in their abstracts. None were available, but they are now available as a plugin.
  3. Next, the UNLP main website was comfortable for users, so the Journals and Congresses websites have been adjusted to look similarly. A comparison may be seen on-screen to show the similarities.
  4. Further, templates are available for users uncomfortable with changing them, but wanting to meet their own needs. OCS and OJS is fully customizable by editing a CSS file. Many CSS styles were made for publishers to choose from because not all had web designers to make changes.
  5. Finally, workshops online have been offered. Users can see an overview to see what the software is capable of and what it won’t do. There are workshops for users and workshops for managers. They have a “sandbox” area where users can try things without affecting the real system.

Note: the voice link with Gonzalo was lost in the last 2 minutes of his presentation. We hope all essential concepts were conveyed.


Blogger’s Thoughts:

The workshops sound tremendously helpful for users and managers. The sandbox feature is very practical and should be of interest to many. A great deal of work has been done to provide leadership in Argentina and its metron of influence. While the audio feed accomplished the task, it would have been good to see them in person to be able to ask questions and benefit from their experience. I was particularly interested in how their design of Celsius-DL provided benefits not found in software as a result of their worldwide search.

In light of the work so many other libraries and institutions are doing to promote the dissemination of scholarly knowledge and research, through either the use of OJS and OCS, one thought seems to echo through the workshops. It is a “cry” for collaboration and communication. From sharing experiences and collaborative problem-solvingj, to establishing some form of index standard are common issues shared by Mexico, the U.K., Canada and others. I hope a leader emerges and this is actualized.

While this is recognized as a technically-themed presentation, I wondered how much use of open journals resulted from their work. While I couldn’t access this information, I was able to view a page on their website that revealed how many “hits” they’ve had at SeDiCi and how many registered users they have. I was surprised that the vast majority originated in the United States! Journal numbers, particularly open journal numbers would be of interest to many at the next conference.

Presentation Link to be provided by conference organizers.

Presenter Links:

Gonzalo Villarreal [The blogger lacked the language to translate, but has listed some found on the SeDiCi repository!]

A Linux publication

Co-authored, in Portugese:  Manuscript Character Recognition Overview of features for the Feature Vector

Marisa Raquel De Giusti

Simulation framework for teaching in modeling and simulation areas

Storage of Simulation and Entities History in discrete models

Article by both outlines the history of their work at La Plata in more detail.

Related Links

PrEBi at La Plata National University

July 10, 2009   Comments Off on Socializing and Disseminating the Academic and Intellectual Creation: Experiences from La Plata National University: The Session Blog

Open Access in Brazil IBICT’s Initiative Drives Country’s Development: The Session Blog

Ramon Martins Sodoma da Fonseca

Ramon Martins Sodoma da Fonseca

Presenter: Ramon Martins Sodoma da Fonseca

July 10, 2009 at 9:30 a.m.

Background

Ramon Fonseca participated in the PKP conference on behalf of the Brazilian government as well as representing the IBICT institution.  In English, IBICT translates as Brazilian Institute of Information in Science and Technology.

Fonseca started his talk out by firstly acknowledging the sponsorship and invitation from the PKP conference; making it possible for him to be here speaking in Vancouver.

Note: I will attempt to use the Google Translate tool for as many links as possible.

Session Overview

IBICT’s main mission is to promote scientific interchange, training and development; contributing to the progress of technology innovation in Brazil.  IBICT strives to expand Brazilian intellectual production nationally and internationally.

One key element of Fonseca’s workshop was to inform the public at large the the Web tools such as blogs, CMS, Google Docs, Twitter, Facebook etc. has allowed numerous opportunities for Brazilians to “have a voice.”  The Web provided the tools and Fonseca claimed Open Access has allowed the exchange of knowledge into and out of Brazil possible.

Some IBICT initiatives:

  • DiCi (1st attempt at open access)
  • Perl (Digital Library Project)
  • TEDE (Online Theses and Dissertation Publishing System)
  • BOTD (Brazilian Digital Library)
  • SEER (meaning OJS in Portuguese)

These software initiatives were launched to specifically to ensure people in Brazil had free access to knowledge.  IBICT also currently provides training and support to national journal editors online which is more cost effective than face to face workshops.

Other Services provided by IBICT:

  • COMUT (Bibliographic Communication Program)
  • CCN (National Collective Catalog Serial Publications)
  • CanalCiencia (Science Communication Channel)

Fonseca also talked about the Social Inclusion Program that gives assistance in accessing online knowledge with children in rural areas of Brazil and technical and infrastructure support to Indigenous communities.

One of the issues that Fonseca discussed was the current importance of maintaining the software and ideas.  Launching the systems and concepts of Open Access is important but these projects need continual restructuring and continued experience for the sustainability of the programs.

The IBICT continues to look for solutions to Open Access and the impact of the Web for making positive and forward thinking decisions regarding technology, system development and strong partnerships for networking the goals and missions of IBICT.

Audience Input

An audience member asked about using MOODLE as a software for an in-house small college class journal.  Fonseca suggested that the college look into OJS because it can adapted into many formats and is a very stable system.

Related Links

Article by Fonseca

DSpace University of New Mexico

Open access book on Open Access

July 10, 2009   Comments Off on Open Access in Brazil IBICT’s Initiative Drives Country’s Development: The Session Blog

XML and Structured Data in the PKP Framework: The Session Blog

Presenter: MJ Suhonos, Session Abstract

July 10, 2009 at 11:00 a.m.

Background

MJ Suhonos is a system developer and librarian with the Public Knowledge Project at Simon Fraser University. He has served as technical editor for a number of Open Access journals, helping them to improve their efficiency and sustainability. More recently, he leads development of PKP’s Lemon8-XML software, as part of their efforts to decrease the cost and effort of electronic publishing, while improving the quality and reach of scholarly communication.

Session Overview

“Lemon8-XML is a web-based application designed to make it easier for non-technical editors and authors to convert scholarly papers from typical word-processor editing formats such as MS-Word .DOC and OpenOffice .ODT, into XML-based publishing layout formats.” (Lemon8-XML).

This was a packed session, 50+ attendees. This technical session attempted to give a fairly non technical overview of the L8X software and its relationship to the PKP software suite and equally importantly to highlight the rich benefits that are provided by using XML workflow and the foundation it provides for the future.

The big question is why use XML workflow. Using XML workflow allows numerous things to be possible. These include interaction with other web services (direct interaction with indexes and better interaction with online reading tools); automatic layout (generate html and/or PDF on the fly); complex citation interaction (forward and reverse linking which allows the discovery of everyone who cited you anywhere on the web; advanced bibliometrics, not just impact measures; resource discovery (universal metadata can find related works; and rich document data allows search engine to be much more effective; the document becomes the metadata (remove separation between article and document so all information is in one place. This is the goal of L8X, to convert articles into structured xml and thus enable these benefits. This is also future proofing as XML makes documents fundamentally open, convertible and preservable. Archiving XML (which is text) is much more flexible than archiving PDF files.

Using XML allows connection and communication to all these systems and means of display. We are also future proofing, as XML will be able to be modified into future formats, as its just text.

Where does this fit within PKP framework? Already being used in OJS (import and export and exposing metadata to OAI harvesters). But the next goal is to apply these benefits to all kinds of scholarly work e.g. journal articles, proceedings, theses, books / monographs. So moving L8X into the PKP web application library will allow all these features to be made available to the whole PKP framework. So that’s the near term future plans for L8X. In the long term, beyond the next few years, the goal is to work on this concept of the doc is the metadata by building support for multiple XML formats in the web application library (WAL) and the merging of annotation, reading tools and comments directly into the article.

The distributed resource-linking diagram at the end of the presentation, some find complex. Essentially, structured metadata is needed to make this a reality, which is to let applications in the publication sphere all talk to each other.

Session Questions
Question: How automatic is automatic into XML for non-technical people? When can I just upload my doc and have it magically turn into XML?
Answer: Probably not ever, but it is semi automatic already. Some tools, like L8X, automate part of this process. Some things can be automated, but some will always require human effort.

Question: Will I be able to use L8X in my applications after this is integrated into the PKP framework?
Answer: We would like to be make L8X available for use after it becomes part of the framework and without requiring the framework. We are considering this for the future.

References and Related Links

Lemon8-XML

Lemon8-XML demo server (login: lemon8  password: xmldoc)

NLM Journal Publishing Tag Set

Open Journal Systems

OpenURL Overview

XML Introduction

July 10, 2009   2 Comments

Revues.org and the Public Knowledge Project: Propositions to Collaborate (remote session): The Session Blog

Presenter: Marin Dacos

July 9, 2009 at 10:00 a.m. SFU Harbour Centre. Rm. 7000*

*Important Note: As this was a remote session, the presenter’s voice was inaudible most of the time due to technical difficulties and constant breaks in the live audio streaming. Therefore, it was difficult to capture parts of the presentation.

Marin DacosMarian Dacos (Source)

Background

Mr. Marin Dacos is a digital humanities specialist who is currently the director of the Centre of Open Electronic Publishing (Cleo) as well as an information systems manager for National Centre for Scientific Research (CNRS) located in France. Furthermore, he is the founder of several initiatives such as Revues.org, Lodel, Calenda, and Hypotheses which will be discussed in the session overview. Mr. Dacos investigates how Revues.org and PKP can be used to improve manuscript management tools and document conversion.

Session Overview

Mr. Martin Dacos initiated the session by providing a background summary on Revues.org. He indicated there are currently 187 members, with 42000 online humanities and social sciences full-text, open-access documents (Session Abstract). He mentioned that approximately ten years ago, systems were centralized and focused on sciences. And since the beginning of Revues.org,  only PDF documents were processed for publishing by converting to extensible markup language (XML).  Later, Lodel (electronic publishing software)  was developed as a central management system (CMS) where the web service could convert word documents to XML. This is around the time when the Public Knowledge Project (PKP)  started up and its focus was to decentralize and provide a more international access point for the publishing of journals and management of conferences through the Open Journal System (OJS). During this time, the two projects, Lodel and PKP, started to converge with two distinct parts and four kinds of services.

(1) The Project Details

The first part of the project, as Mr. Dacos described it consists of using PKP  to develop a manuscript management tool to monitor the workflow through OJS. There is a need to create a new interface for users and make it more human-friendly for interaction in order to allow for the dissemination of documents. This portion of the project also investigates the possibility of connecting Lodel and OJS so both systems can use the system jointly. Next, Mr. Dacos explained the second part of the project which deals with document conversion called OTX – which will convert for example RTF to XML. This parallels PKPs development and there is the possibility of sharing information on this creation.

(2) Services

Revues.org offers various kinds of services to allow for the dissemination and communication of scholarly material and other information such as upcoming events. One of the services presented by Revues is Calenda which is claimed to be the largest French calendar system for the social sciences and humanities. This calender service is important because it disseminates information such as upcoming scientific events to the rest of it’s audience. This communication tool is crucial in bringing members of various online communities together to participate in ‘study days,’ lectures, workshops, seminars, symposiums, and share their papers. Another valuable service offered by Revues.org is Hypotheses which is a platform for research documents. This is a free service which allows researchers, scientists, engineers and other professionals to post their experiences on a particular topic or phenomenon for sharing with a wider audience. One can upload a blog, field notes, newsletters, diary inserts, reviews on certain topics, or even a book for publishing. A third service offered by Revues.org is a monthly newsletter called La Lettre de Revues.org. This newsletter connects the Revues.org community together by showcasing various pieces of information. For instance, new members who have recently joined are profiled and new online documents are highlighted for it’s subscribers to read.

The remainder of Mr. Marin Dacos’ talk focused on Lemon8 and OTX which was difficult to interpret due to technical issues.

Questions from the audience asked at Mr. Dacos’ session:

It was difficult to get an audio connection with Mr. Dacos due to technical difficulties, therefore questions were not asked.

Related Links

References

Dacos, M. (2009). Revues.org and the public knowledge project: propositions to collaborate. PKP Scholarly Publishing Conference 2009. Retrieved 2009-07-09, from http://pkp.sfu.ca/ocs/pkp/index.php/pkp2009/pkp2009/paper/view/208

July 10, 2009   Comments Off on Revues.org and the Public Knowledge Project: Propositions to Collaborate (remote session): The Session Blog

The new Érudit publishing platform: The Session Blog

Presenter: Martin Boucher

July 9, 2009 at 9:30 a.m. SFU Harbour Centre. Rm 7000

Martin Boucher (picture taken by Pam Gill)

Martin Boucher (picture taken by Pam Gill)

Background

Mr. Martin Boucher is the Assistant Director for the Centre d’édition numérique/ Digital Publishing Centre located at the University of Montreal, located in Montreal, Quebec. This centre along with the library at the University of Laval are the sites of the Erudit publishing locations which serve as a bridge to the Open Journal Systems (OJS). Erudit focuses on the promotion and dissemination of research similar to the Open Access Press (part of the PKP 2009 conference).

Session Overview

Mr. Martin Boucher highlighted the features and implications of a new publishing platform called Erudit (Session Abstract). He focused on sharing the capabilities of the Erudit publishing platform by first providing a brief historical overview of the organization, describing the publishing process, then introducing Erudit, and concluding with the benefits of such a platform.

(1) Historical Overview

Mr. Boucher started the session by pointing out that Erudit is a non-profit, multi-institutional publishing platform founded in 1998.  This platform, based in Quebec, provides an independent research publication service which consists of access to various types of documents in the humanities and social sciences fields to the universities. Erudit also encouraged the development of Synergies which is a similar platform but targets a more mainstream audience since it is published in English. Some facts about Erudit:

  • International standards are followed
  • Publishes over 50000 current and back-dated articles
  • Offers management services, publishing, and subscriptions
  • 90% of the downloads are free
  • Have over 1 million visits per month

(2) Publishing Process

The description of the publishing process constituted at least a third of the presentation time. Mr. Boucher felt it necessary to take the time to describe to the audience the details involved so it would be easier to compare the similarities and differences between the new and old versions of the system. To begin, Mr. Boucher indicated the publishing process accepted only journals based on extensible markup language (XML), and various input sources (In Design, QuarkXPress, Open Office, Word, RTF). Also, he pointed out that there are no peer-reviews, in fact: only the final documents are considered to be a part of the collection. It should be pointed out that these documents meet high quality standards as they are expected to be peer-reviewed before submitting to the publishing platform. As of yet, Erudit does not have the software to assist in a peer-review type of process. The belief of the Erudit community is to provide quick digital dissemination of the articles. This complicated, lengthy process is made possible by a team of three to four qualified technicians, one coordinator, and one analyst, all of whom ensure a smooth transition of the documents into the virtual domain. The publishing process consists of five key steps as outlined in Figure 1. Mr. Boucher elaborated on the importance of the analysis step. He went on to outline the three steps of the manual semantic analysis. The first consists of manual and automated tagging where detailed XML tagging is only for XHTML, and less tagging is done on PDF files. The second step consists of the automated production of XML files for dissemination. And lastly, a rigorous quality assurance by the technicians prior to dissemination sums up the analysis step of the publishing process.

Figure 1: Publishing Process (image created by Pam Gill)

Figure 1: The publishing process (image created by Pam Gill)

(3) New Erudit Platform

Once the publishing process was described, an illustration of the new Erudit platform was revealed. Mr. Boucher indicated there is now increased support for journal articles through the Erudit Article 3.0 XML schema (see Figure 2). Further, there is support for additional scholarly genres such as books, proceedings, and even online courses (a recent request). It should be mentioned here that some of these other forms of documents are still in the experimental stages such as digitizing books. In addition, there is continued support for other XML input/output formats to ensure preservation and interoperability such as with the Open Archives Initiative (OAI), Google Scholar, and indexing databases.

Figure 2: The new Erudit support system (image created by Pam Gill)

Figure 2: The new Erudit support system (image created by Pam Gill)

(4) Benefits

To conclude his presentation on Erudit, Mr. Boucher explained the advantages of incorporating such a system by mentioning particular benefits of interest:

  • Purely Java-based
  • User-friendly because they have a universal set of tools inside the applications which makes it easier for the support technicians to troubleshoot and work with
  • Supports plug-ins and extensions
  • All scholarly genres are supported
  • The process is simpler to follow
  • An increase in the quality of data is noted
  • The decrease in production time is evident
  • There is less software involved

Mr. Martin Boucher hinted that the beta version of Erudit was to release in Fall 2009.

Questions from the audience asked at Mr. Boucher’s session:

  1. Question: For the open access subscription of readership which consists of a vast collection, are statistics being collected? Answer: Not sure.
  2. Question: Will the beta version of the publishing platform be released to everyone for bug reporting, testing, or move internally? Answer: Not sure if there will be public access. But it is a good idea to try the beta platform.
  3. Question: Are you considering using the manuscript coverage for the Synergies launch? Answer: The new platform is creatively tight to what we are doing, and it is really close, with Synergies in mind.
  4. Question: In a production crisis, are journal editors with you until the end of the process? Answer: They are there at the beginning of the process. They give material, but we do our own quality assurance process and then we release to the journal, however it is our own control. Also, the editors cannot see the work in the process such as the metadata, thought we do exchange information by emails.
  5. Question: Has the provincial government been generous in funding? Answer: The journals had to publish in other platforms. There is a special grant for that. It is easier for us with that granting repository for pre-prints, documents or data section of the platforms which serve as an agent for them. Yet, Erudit is not considered by the government, although we are trying to get grants from the government. Currently to maintain the platform we only have money for basic management. In order to continue developing platforms (such as Synergy), to get support from the government is difficult.
  6. Question: How do the sales work for the two platforms? Answer: If you buy it, you will have all the content and access increases.
  7. Question: How is it passed to the publisher? Answer: The money goes to the journals, keep only a small amount for internal management since we are a non-profit society.
  8. Question: Could you describe the current workflow and time required to publish one article? Answer: It depends on the article. If we are publishing an article that has no fine grain XML tagging or it is text from a PDF, then it requires less time for us to get it out. It depends on the quality of the article and the associated graphics, tables, size etc. We publish an issue at a time. It takes say two days to get an article published.

Related Links

University of Montreal receives $14M  for innovation (news article)

Contact the University of Montreal or the University of Laval libraries for more information on Erudit.

References

Boucher, M. (2009). The new erudit publishing platform. PKP Scholarly Publishing Conference 2009. Retrieved 2009-07-09, from http://pkp.sfu.ca/ocs/pkp/index.php/pkp2009/pkp2009/paper/view/182

July 10, 2009   2 Comments