Re: Splitting a large file of MARC records into smaller files

2010-01-25 Thread Saiful Amin
I also recommend using MARC::Batch. Attached is a simple script I wrote for myself. Saiful Amin +91-9343826438 On Mon, Jan 25, 2010 at 8:33 PM, Robert Fox wrote: > Assuming that memory won't be an issue, you could use MARC::Batch to > read in the record set and print out seperate files where y

Re: Splitting a large file of MARC records into smaller files

2010-01-25 Thread Robert Fox
Assuming that memory won't be an issue, you could use MARC::Batch to read in the record set and print out seperate files where you split on X amount of records. You would have an iterative loop loading each record from the large batch, and a counter variable that would get reset after X amo

Re: Splitting a large file of MARC records into smaller files

2010-01-25 Thread Sébastien Hinderer
Hi, The yaz-marcdump utility may be what you are looking for. See for instance options -s and -C. hth, Shérab.

Re: Marc::XML with MARC21

2010-01-25 Thread Ed Summers
Hi Michele: I copied and pasted the XML from your email and ran it through a simple test script (both attached) and the record seemed to be parsed ok. What do you see if you run the attached test.pl? //Ed test.pl Description: Binary data http://www.loc.gov/MARC21/slim"; xmlns:xsi="http://ww

RE: Splitting a large file of MARC records into smaller files

2010-01-25 Thread Walker, David
> yaz-marcdump allows you to break a > marcfile into chunks of x-records +1 --Dave == David Walker Library Web Services Manager California State University http://xerxes.calstate.edu From: Colin Campbell [colin.campb...@ptfs-europe.com]

Re: Splitting a large file of MARC records into smaller files

2010-01-25 Thread Ashley Sanders
Jennifer, I am working with files of MARC records that are over a million records each. I'd like to split them down into smaller chunks, preferably using a command line. MARCedit works, but is slow and made for the desktop. I've looked around and haven't found anything truly useful- Endeavor'

RE: Splitting a large file of MARC records into smaller files

2010-01-25 Thread Smith,Devon
This isn't a perl solution, but it may work for you. You can use the unix split command to split a file into several other files with the same number of lines each. For that to work, you'll first have to use tr to convert the ^] record separators into newlines. Then use tr to convert them all back

Re: Splitting a large file of MARC records into smaller files

2010-01-25 Thread Colin Campbell
On 25/01/10 14:48, Nolte, Jennifer wrote: Hello- I am working with files of MARC records that are over a million records each. I'd like to split them down into smaller chunks, preferably using a command line. MARCedit works, but is slow and made for the desktop. I've looked around and haven't

RE: Splitting a large file of MARC records into smaller files

2010-01-25 Thread Houghton,Andrew
> From: Nolte, Jennifer [mailto:jennifer.no...@yale.edu] > Sent: Monday, January 25, 2010 09:48 AM > To: perl4lib@perl.org > Subject: Splitting a large file of MARC records into smaller files > > Hello- > > I am working with files of MARC records that are over a million records > each. I'd like t

Re: Splitting a large file of MARC records into smaller files

2010-01-25 Thread Emmanuel Di Pretoro
Hi, A long time ago, I've written the following : --- snippet --- #!/usr/bin/env perl use strict; use warnings; use MARC::File::USMARC; use MARC::Record; use Getopt::Long; my $config = { output => 'input' }; GetOptions($config, 'input=s', 'chunk=s', 'output=s', 'max=s'); if (not exists $con

Splitting a large file of MARC records into smaller files

2010-01-25 Thread Nolte, Jennifer
Hello- I am working with files of MARC records that are over a million records each. I'd like to split them down into smaller chunks, preferably using a command line. MARCedit works, but is slow and made for the desktop. I've looked around and haven't found anything truly useful- Endeavor's MAR

Re: Marc::XML with MARC21

2010-01-25 Thread Jon Gorman
> > my $file = MARC::Record->new_from_xml($marc->serialize(),"UTF-8","MARC21"); >        $epdata = $plugin->EPrints::Plugin::Import::MARC::convert_input( > $file ); > > and here come troubles: only few metadatas will be interpreted > correctly, losing a lot of datas. Ummm, so what metdata makes it

Marc::XML with MARC21

2010-01-25 Thread Michele Pinassi
-BEGIN PGP SIGNED MESSAGE- Hash: SHA1 Hi all, im'm working on a perl plugin for EPrints that let user importing from Aleph simply using system id. It use Aleph OAI-PMH service that export metadata in MARC21 format: http://www.openarchives.org/OAI/2.0/ http://www.openarchives.org/OAI/2.0/O