Friday, November 03, 2006

Web Reference Database- 0.9.0 Released


Refbase is a web-based bibliographic manager for scientific literature, references and citations. A new release of the refbase package is available at:

http://prdownloads.sourceforge.net/refbase/refbase-0.9.0.tar.gz?download

This release offers major function enhancements and bugfixes. Batch import from various bibliographic formats (including BibTeX, Endnote, RIS, ISI and MODS XML) is now supported, as is import from a PubMed ID. An OpenDocument Spreadsheet for use with OpenOffice.org can be exported and formatted citation lists can be generated as HTML, RTF, PDF, or LaTeX. A SRU/W service as well as support for unAPI, OpenURL and COinS metadata have been added. These allow the data to be used by the next generation of bibliographic clients. A new command line client is also included.

An overview of the main features of this package can be found at:

http://refbase.sourceforge.net/features.html

ARCHIMEDE(2.0 version) : A canadian software solution for institutional repositories

Archimède is an electronic document management system, designed and developed at the library of the Université Laval.

Here’s a list of new features for the upcoming Archimede 2.0 version:

  • Fine-grained security
  • Automatic versioning of content
  • Supports for locks (for exclusive content editing)
  • Support for custom workflows (editable at run-time)
  • Web-Dav + DeltaV interface
  • Custom metadata formats (run-time addition)
  • Improved Web interface
  • Support for portal integration (JSR-168 portlets)
  • Fully implements the Java Content Repository standard (JSR-170). Archimede is built on top of the Jackrabbit reference implementation.
  • Pluggable authentication/authorization module that uses Java Authentication and Authorization Service. (For example, LDAP authentication can be easily added.)
  • Flexible event notification (new content, content modification, etc.)
open source distribution of Archimede is now available.

Digital Preservation Recorder (DPR)

Digital Preservation Recorder (DPR) has been developed by the National Archives
of Australia to manage a digital preservation workflow.

Features of the DPR
* Cross platform – uses a Java Runtime Environment.
* Metadata recording.
* User authentication and profiling.
* Multiple concurrent users.
* Detailed reporting.
Preservation

The carrying device is connected to the Preservation network
and the individual data files are converted to XML via the open
source application Xena. The XML files created by Xena are recorded on
a second carrying device for transfer to the Digital Repository.

Downloading DPR

http://sourceforge.net/project/showfiles.php?group_id=116934&package_id=127214

How to start DPR:
There are two ways to start DPR. Either may be effective depending on the Java Runtime Environment installation on a particular computer.

The simplest way is to open a file manager via Windows Explorer or My Computer and navigate to the dprclient.jar file. Double click the jar file and wait a few moments for
the Java Virtual Machine to start up and the DPR login to display.

On some computers, it may be necessary to start DPR manually. Open a command prompt (Start, Run, cmd, OK) and navigate to the directory containing the dprclient.jar file.

Type:

java -jar dprclient.jar

The Java Virtual Machine should start in a few moments and the command prompt window may display some startup messages. The DPR user interface will follow and DPR is ready for use.

Technical contact: digipres@naa.gov.au

Source:http:http://sourceforge.net/projects/dpr

GeoNetwork:web based Geographic Metadata Catalog System



GeoNetwork is a web based Geographic Metadata Catalog System developed by FAO-UN, WFP-UN and UNEP. The system implements the ISO 19115 Geographic Metadata and ISO 23950 (Z39.50) standards. Technical details: Java, XML/XSL, JDBC compliant, Multi-lingual .

Data Crow




Data Crow : Software tool for manage list of movie & video, book, software (games & program) and music cataloguer / database. Uses online web services (amazon, imdb, musicbrainz, freedb). Skinnable, highly customizable, creates reports. Supports: DVD, many audio & video file formats and Audio CD's . Search for your titles in the Amazon database. Amazon has a huge database holding information on any kind of book. Data Crow retrieves descriptions, information about the author, year of release, average user rating, a front cover picture and more! Just like with all the other modules you can easily retrieve this information and ofcourse change it. The information can be exported to a pdf or html file for nice print-outs

November issue of the SPARC Open Access Newsletter

"This issue reviews the eleven (count 'em) open-access mandates adopted or proposed in October and takes an opening reconnaissance into the largely unexplored territory of no-fee OA journals. The Top Stories section takes a brief look at the Citizendium project, the continuing division within the American Anthropological Association over OA and FRPAA, the new NESLi2 model license that allows self-archiving, the new JISC-SURF model license that lets authors retain key rights, and Google Custom Search. "

"Web reaches new milestone: 100 million sites"

"There were just 18,000 Web sites when Netcraft, based in Bath, England, began keeping track in August of 1995. It took until May of 2004 to reach the 50 million milestone; then only 30 more months to hit 100 million, late in the month of October 2006.Netcraft says the United States, Germany, China, South Korea and Japan show the greatest Web site growing spurts.
When the Web was started, it was started as a mechanism for sharing high energy particle physics data," said Professor Rebecca Grinter of Georgia Tech's College of Computing. The creator of that Web site, Tim Berners-Lee, wanted experts to be able to share data on particle smashing, even if they weren't at CERN in Switzerland where he was doing research. CERN, in Geneva, is the European Organization for Nuclear Research.Research facilities and universities soon started seeing benefits of this new tool for things as lofty as nuclear physics and as mundane as sharing restaurant recommendations.......................More"


Thursday, November 02, 2006

Google Brain

Google Next ..........

INTELLIGENT MEDIA MANAGER

PhiloLogic™




PhiloLogic™ is the primary full-text search, retrieval and analysis tool developed by the ARTFL Project and the Digital Library Development Center (DLDC) at the University of Chicago. This is a Free Software implementation of PhiloLogic for large TEI-Lite document collections. The wide array of XML data specifications and the recent deployment of basic XML processing tools provides an important opportunity for the collaborative development of higher-level, interoperable tools for Humanities Computing applications.


  • light, fast, robust, extensively used and tested
  • few dependencies, basic installation almost wholly self-contained
  • out of the box operation with many configuration options
  • TEI-Lite XML/SGML (and variants such as MEP and CES) with Unicode support
  • support for plaintext, Dublin Core/HTML, and DocBook
  • MySQL back-end for bibliographic searching
  • optional XML-aware or non-XML bibliographic loaders
  • interoperability across certain systems
  • fault tolerant
  • open source


  • Open Source Tool: Web Curator Tool : 1.1 GA Released


    1.1 GA 21 September 2006 Release notes, known issues, and changelog

    The Web Curator Tool (WCT) is a tool for managing the selective web harvesting process. It is designed for use in libraries and other collecting organisations, and supports collection by non-technical users while still allowing complete control of the web harvesting process.


    Web Curator Tool as an open-source project.

    New features

    • ReferenceNumber field for Target, Group and TargetInstance objects
    • FileReference field for Permission objects
    • Fields for recording Selection information in Target objects
    • HarvestType field for Target objects
    • Add a textbox labelled ?Name? to the Target Instance Search Form
    • ProfileNote field for Target objects
    • Basic descriptive (Dublin Core) metadata
    • Type field for Group objects
    • Sticky search forms: reset button needed [SF 1531646]
    The tool's workflow encompasses the following tasks:


    * Harvest Authorisation: seeking and recording permission to harvest web material, and to make it accessible to the general public.

    * Selection and scoping: determining what material should be harvested, be it a
    web site, a web page, a partial web site, a group (or collection) of web sites,
    or any combination of these.

    * Scheduling: determining when a harvest should occur, and when it should be
    repeated.

    * Description: describing harvests with basic Dublin Core metadata, and other
    specialized fields (or a by a providing a reference to an external catalogue).

    * Harvesting: the Web Curator Tool will download the selected web material at
    the appointed time using the Internet Archive's Heritrix web crawler -- each
    installation can have multiple harvesters on different machines, each which can
    perform several harvests simultaneously.

    * Quality Review: tools are provided for making sure the harvest worked as
    expected, and correcting simple harvest errors.

    * Endorsing and submitting: if the harvest was a success, it is endorsed then
    submitted to an external digital archive.

    Tuesday, October 31, 2006

    Semantic Personal Digital Library (SPDL)


    The Semantic Personal Digital Library (SPDL) is an open source Java project helping user to organize and retrieve documents in PDF format. It is a software of Personal Information Management (PIM) that works with technologies of Semantic Web. User can insert/edit informations on documents like author, title, description, subject and so on. These informations are stored as RDF (Resource Description Framework) and use standard properties like those defined in Dublin Core metadata set. User can also build some simple ontologies helping user to classify documents by the subjects they are about or other free classifications. Retrieving documents can be done searching on content and metadata values or browsing classification schemes. SPDL uses Lucene index and SPARQL to search documents. SPDL is a personal accademic project to help user in personal information management, but it's also an example of how technologies of Semantic Web can be used in real project where information has to be organized and retrieved.

    • Acquire items into PIM system
    • Organize items in personal collection or in classification schemes
    • Maintain items in the system (edit metadata information, delete items...)
    • Retrieve items searching for metadata values or browsing classification schemes

    SPDL actually manage only PDF documents as digital items, but support for other items will be added in the future (word documents, power point, etc.)


    Download: full release download.



    Personalized Information Services : My Library Projects

    My Gateway
    University of Washington Libraries

    My Gateway debuted in the fall of 1998. Includes remote authentication and access to circulation information services. As of January 1999 about 1500 users had created My Gateway pages.
    http://www.lib.washington.edu/resource/help/MyGateway.html

    BraryDog
    Public Library of Charlotte & Mecklenburg County

    "Your Homework Help and Web Companion." Written in ASP, this is a noteworthy example from a public library. Great graphics and appealing promotion as an integrated tool for education. Includes The Brarydog Rap.
    http://www.brarydog.net

    My Library
    California Polytechnic State University Library

    This personalized library webpage service is integrated into document delivery and is targeted to graduate students and faculty.
    http://www.lib.calpoly.edu/mylib/cgi-bin/index.cgi

    My Library
    Cornell University Library

    Consists of MyLinks, a tool for collecting and organizing resources for private use by a patron, and MyUpdates, a tool to help scholars stay informed of new resources provided by the library. For more information see the April 2000 article in D-Lib Magazine.
    http://mylibrary.cornell.edu/

    The Portal Formerly Known As MyLibrary@NCState
    Hosted at Notre Dame University Libraries

    Offers subject specific groupings of content, updates, and remote authentication, extensive descriptions about the architecture, planning, and administration of the system. Includes guest access so you can take it for a test drive. Source code is available for download at http://dewey.library.nd.edu/mylibrary/.

    A selected list of libraries employing the NC State code:

    MyWelch
    Johns Hopkins Welch Medical Library

    Written in Cold Fusion. When you register, you select your areas of interest, and then appropriate resources are selected as your defaults (you can modify them later). Includes one click access to other services such as e-mail and borrower information, news feeds, links to online reference, and a personal note pad. Includes guest access. Version 3.0 was released April 2004 with expanded features including a UMLS Concept Search, Tab Interface, Web File Server, and Weather Room.
    https://mywelch.welch.jhmi.edu/

    Personalized Eccles
    University of Utah, Health Sciences Center

    Based on the library's homepage design, this is an interesting example of the use of personalization within a specialized library. The source code is available from Wayne Peay.
    http://medstat.med.utah.edu/personalize/welcome.html

    My Library@UT Southwestern Medical Center Library
    University of Texas Southwestern Medical Center

    Written with Cold Fusion, users can add choices from different modules including direct links to e-journals. Includes an e-mail alertint feature..
    http://www2.utsouthwestern.edu/cfdocs/library/mylibrary/myliblogin.html

    my.library
    University of Toronto Libraries

    Written with Cold Fusion..
    http://www.library.utoronto.ca/mylibrary/

    MyLibrary
    Christchurch College of Education, ChristChurch, New Zealand

    Written with PHP..
    http://mylibrary.cce.ac.nz/index.php?Action=login

    My Library @ CDL
    California Digital Library

    Written with Cold Fusion..
    http://mylibrary.cdlib.org/

    myLibrary
    Mississippi State University Libraries

    Written with ASP. Provides guest access.
    http://library.msstate.edu/mylibrary/index.asp

    Selected University Portal Projects

    My UCLA
    University of California Los Angeles

    Since its debut in October 1997, over 33,000 different students have successfully created pages for a total of over 1.2 million distinct authenticated sessions. My UCLA includes access to class schedules, campus information and web based email (using student selectable pop/smtp servers) but without a unique library component. Includes a guest page and technical documentation.
    http://my.ucla.edu

    uPortal
    JA-SIG

    "uPortal is an open-standard effort using Java, XML, JSP and J2EE. It is a collaborative development project with the effort shared among several of the JA-SIG member institutions. You may download uPortal and use it on your site at no cost. "
    http://mis105.mis.udel.edu/ja-sig/uportal/

    FOR updation see this link :http://www.library.vcu.edu/mylibrary/cil99.html

    uPortal

    uPortal is an open-standard effort using Java, XML, JSP and J2EE. It is a collaborative development project with the effort shared among several of the JA-SIG member institutions. You may download uPortal and use it on your site at no cost.

    download link: http://uportal.org/download.html

    Institutions implementing uPortal

    For additions and updates please contact: webmaster@ja-sig.org

    • ALBI (France)
    • Antonine University
    • BCcampus
    • Brooklyn College
    • Carl Sandburg College
    • Centre De Recherche Informatique Montreal
    • Colorado State University
    • Dallas Baptist University
    • eXtension Initiative
    • Florida State University
    • Georgia Institute of Technology
    • Hofstra University
    • INPL (France)
    • INPT (France)
    • IState (Iowa State University)
    • Ireland Integrated Learn. Environ.
    • IUFM de Bretagne
    • Kwantlen University College
    • Lawrence Berkeley National Lab
    • Lethbridge Community College
    • London Metropolitan University
    • Marietta College
    • Mississippi State University
    • Muhlenberg College
    • New Mexico State University
    • NHMCCD
    • NSF National Middleware Infrastructure
    • Pontificia Universidad Catolica de Chile
    • Red Universiaria Nacional de Chile
    • Rockefeller University
    • Sabanci University
    • San Joaquinn Valley College
    • Swiss Fed Institute of Tech, Zurich
    • Trinity University
    • Tulare County Office of Ed
    • UC Merced
    • UC Santa Cruz New Teacher Center
    • Université de Valenciennes
    • University College of Fraser Valley
    • University of Birmingham, UK
    • University of Bremen
    • University of Bristol
    • University of Edinburgh
    • University of Franche Comte
    • University of Havre
    • University of Helsinki
    • University of Latvia
    • University of Littoral
    • University of London Royal Hollowal
    • University of Miami
    • University of Nancy 1
    • University of Nevada, College of Ed
    • University of Otago, New Zealand
    • University of Oxford
    • University of Rennes 1
    • University of Rennes 2
    • University of Texas at Dallas
    • University of Toronto at Mississauga
    • University of Toulouse 3
    • University of KwaZulu-Natal
    • University of Versailles
    • University of West Florida
    • Valparaiso University
    • Washington Lee School of Law
    • Webster University

    Simple Digital Library Interoperability Protocol (SDLIP-Core)



    SDLIP: The Simple Digital Library Interoperability Protocol (SDLIP; pronounced S-D-Lip) is a protocol for integrating multiple, heterogeneous information sources. It was developed jointly by Stanford University, UC Berkeley, UC Santa Barbara, the San Diego Supercomputer Center, and the California Digital Library Project. Clients use SDLIP to request searches to be performed over information sources.

    CAUT Copyright Conference-Notes

    The Canadian Association of University Teachers hosted a conference this past weekend (October 27-29) titled “Controlling Intellectual Property: The Academic Community and the Future of Knowledge.” it is published in digitalelectric


    Michael Geist: Copyright and Control in the Digital Age
    Cancopy law, CMEC's Copyright Matters!; Captain Copyright (Access copyright)
    -these are very similar - you don't see user rights in there; it's as if SCC decisions never happened
    -education needs a new vision of copyright
    -The good news story: blogs, myspace, postsecret, flikr,
    -knowledge sharing, wikipedia, wikinews, project Gutenberg, MIT open courseware,
    -the incentive s are different from traditional copyright incentives
    -CIHR policy on open access, nature peer review online trial, Alouette Canada - open digitization project
    -all kinds of open source software
    -SCC reshaping of view of copyright from Theberge, CCH,

    The bad news:
    C-60 will be revived in worse form
    -tepid gains for education
    -WIPO & problems for security researchers
    -insufficiency of fair dealing - Telus has been talking about this
    -StatsCan hadn't taken open licensing as a category in its innovation analysis bulletin
    -education exception might drive a wedge between AUCC and students, librarians

    What to do?
    -not focus on an internet exception
    -DRM - this is the main issue on which education will lose money
    -open access models
    -open licensing
    -expanded fair dealing
    -use CCH
    -drop internet exception
    -focus on digitization - build national digital library - do more than Alouette - Google concerned about getting involved as result of our laws
    -get rid of crown copyright
    -freeze copyright terms - there is already a consultation document on extending it
    -focus on contractual limits - contracts must reflect principles....




    Monday, October 30, 2006

    "The Google Storey"

    Cited Reference Searching

    Specialized Custom Search Engine for Indian libraries and librarians

    Specialized Custom Search Engine for Indian libraries and librarians

    I would be incredibly grateful if someone is willing to help me with this. I can be reached at my email: hbmallikarjuna@gmail.com. Once again, thank you for all of your help and dedication. I need your comments or suggestions on how to improve the search engine and make it more useful, When leaving comments, please let us know who you are and what you do, as this will help us put the feedback in context. If you ever have any problems, questions, comments, or suggestions, please contact. hbmallikarjuna@gmail.com.
    Home page:

    DPubS (Digital Publishing System)

    DPubS (Digital Publishing System) is an open-source software system designed to enable the organization, presentation, and delivery of scholarly journals, monographs, conference proceedings, and other common and evolving means of academic discourse. DPubS was conceived by Cornell University Library to aid colleges and universities in managing and disseminating the intellectual discoveries and writing of scholars and researchers.

    For download:http://sourceforge.net/projects/dpubs

    DPubS has the following minimum requirements:

    • Hardware
      • Tested DPubS on Solaris 9 (sparc) Solaris 10 (sparc) and Linux (Red Hat FC4 for x86).
      • DPubS should work fine anywhere you can run Apache/mod_perl.
    • Perl 5.8+

    • Recommend perl 5.8+ because of the unicode support that was adding with 5.8.
    • Apache/mod_perl
      • Apache (1.3.x) and mod_perl (1.x) or Apache (2.x) and mod_perl (2.x).
    • Java
      • If Lucene search engine is used, Java Runtime Environment (JRE) 1.4.2 or greater is required.
    • Storage Space
      • The disk space needed for DPubS will depend primarily on the amount and type of content that you are going to publish.
      • The DPubS source code itself requires less than 10 MB of space.
    • RAM
      • A minimum of 256MB of memory is recommended for DPubS.



    Administrator Guide

    http://wiki.library.cornell.edu/wiki/display/dpubs/AdminFunctions

    Presentation DPubS:

    This presentation will provide information about DPubS, the Digital Publishing System project that Penn State is working on with Cornell University.

    http://wiki.library.cornell.edu/wiki/display/dpubs/Home

    Download the presentation (2.7 MB PPT file)


    Eric G. Ferrin discussing DPubS, Digital Publishing System.





    Digital Reference Services Bibliography

    Digital Reference Services Bibliography by Bernie Sloan

    More than 700 items listed in this bibliography relate to the topic of online or virtual or digital reference services, i.e., the provision of reference services, involving collaboration between library user and librarian, in a computer-based medium. These services can utilize various media, including e-mail, Web forms, chat, video, Web customer call center software, Voice over Internet Protocol (VoIP), etc.

    Nutch 0.8.1 is now available


    Nutch is open source web-search software. It builds on Lucene Java, adding web-specifics, such as a crawler, a link-graph database, parsers for HTML and other document formats, etc.

    For more information about Nutch, please see the Nutch wiki.




    for DOWNLOAD: http://www.apache.org/dyn/closer.cgi/lucene/nutch/


    Changes: 0.8 to 0.8.1

    1. Changed log4j confiquration to log to stdout on commandline
    tools (siren)

    2. NUTCH-260 - Updated hadoop.jar to contain patch from HADOOP-387
    (siren)

    3. NUTCH-344 - Fix for thread blocking issue (Greg Kim via siren)

    4. Optionally skip pages with abnormally large Crawl-Delay values
    (Dennis Kubes via ab)

    5. Fix incorrect calculation of max and min scores in readdb -stats
    (Chris Schneider via ab)

    6. NUTCH-348 - Fix Generator to select highest scoring pages (Chris
    Schneider and Stefan Groschupf via ab)

    7. NUTCH-338 - Remove the text parser as an option for parsing PDF files
    in parse-plugins.xml (Chris A. Mattmann via siren)

    8. NUTCH-105 - Network error during robots.txt fetch causes file to
    beignored (Greg Kim via siren)

    9. Use a CombiningCollector when calculating readdb -stats. This
    drastically reduces the size of intermediate data, resulting in
    significant speed-ups for large databases (ab)

    10. NUTCH-332 - Fix doubling score caused by links to self (Stefan
    Groschupf via ab)

    11. NUTCH-336 - Differentiate between newly discovered pages and newly
    injected pages (Chris Schneider via ab) NOTE: this changes the
    scoring API, filter implementations need to be updated.

    12. NUTCH-337 - Fetcher ignores the fetcher.parse value (Stefan Groschupf
    via ab)

    13. NUTCH-350 - Urls blocked by http.max.delays incorrectly marked as GONE
    (Stefan Groschupf via ab)

    Koha 2.2.6 (for windows)has been released.
















    Koha 2.2.6 has been released and is available for download here


    Koha is a full-featured open-source ILS. Developed initially in New Zealand by Katipo Communications Ltd and first deployed in January of 2000 for Horowhenua Library Trust, it is currently maintained by a team of software providers and library technology staff from around the globe.


    Koha V2.2.6 Stable Release for Windows
    This is a preliminary test build of the 2.2.6 install package for Windows. It does not include any data. It also lacks several perl modules that are required by version 2.2.6. A new compile that corrects these deficiencies should be available in a few days.
    (Install build R0 last updated October 26, 2006 - 8.7 MB)