-
Notifications
You must be signed in to change notification settings - Fork 36
FAQ
laptop:~/research/source/scraperwiki-com/uk-offshore-oil-wells/version/2011-Jan-24$ ./convert-uk-offshore-oil-wells.sh
--------------------------------------------------------------------------------
uk-offshore-oil-wells.csv
convert-uk-offshore-oil-wells.sh converting newlines of source/uk-offshore-oil-wells.csv
10720 rows in source/uk-offshore-oil-wells.csv
RAW CONVERSION
Error occurred during initialization of VM
Could not reserve enough space for object heap
Could not create the Java virtual machine.
cat: automatic/uk-offshore-oil-wells.csv.raw.ttl: No such file or directory
convert.sh done
convert-aggregate.sh delaying publishing until an enhancement is available.
To publish with only raw, set CSV2RDF4LOD_PUBLISH_DELAY_UNTIL_ENHANCED="false" in $CSV2RDF4LOD_HOME/source-me.sh.
To publish raw with enhanced, add enhancement to manual/uk-offshore-oil-wells.csv.e1.params.ttl and rerun convert...wells.sh
To force publishing now, run publish/bin/publish.sh
===========================================================================================
By default, csv2rdf4lod-automation requests 3GB of memory for csv2rdf4lod.
To reduce that, in your source-me.sh, change:
export CSV2RDF4LOD_CONVERTER=""
to:
export CSV2RDF4LOD_CONVERTER="java edu.rpi.tw.data.csv.CSVtoRDF"
I think there is 0th col in your example
conversion:enhance [
ov:csvCol 0;
conversion:predicate rdfs:seeAlso;
conversion:object
<https://www.og.decc.gov.uk/pls/wons/wdep0100.qryWell>;
];
The first column is column 1. We're humans, not computers. Any enhancement with ov:csvCol 0 is referring to the row.
See the descriptions at conversion:Enhancement and conversion:predicate.
- Dependencies are listed in Installing csv2rdf4lod automation - complete
bash-3.2$ java edu.rpi.tw.data.csv.CSVtoRDF --version
CSVtoRDF: version 2011-Jan-23
bash-3.2$ java edu.rpi.tw.data.csv.CSVtoRDF --version
CSVtoRDF: version 2011-Apr-02
DATE FAILED: "1989-10-02" !~ 0 templates @ :thing_2 :completion_date
- Well, I guess just Tetherless World :-/
- Please let us know if you know of any other adoptions.
- See List of SPARQL endpoints containing datasets produced by csv2rdf4lod
$ java edu.rpi.tw.data.csv.impl.CSVHeaders
usage: CSVHeaders <file> [--header-line headerLineNumber] [--delimiter delimiter]
java edu.rpi.tw.data.csv.impl.CSVHeaders manual/current.csv
If the headers are the second line and the file is pipe delimited:
java edu.rpi.tw.data.csv.impl.CSVHeaders manual/current.csv --header-line 2 --delimiter \|
java edu.rpi.tw.data.csv.impl.CSVHeaders source/gene2ensembl --header-line 1 --delimiter '\t'
When in a conversion cockpit, both manual/x.e1.params.ttl and manual/x.global.e1.params.ttl can appear (where x is some file name). The most common is manual/x.e1.params.ttl and is created by cr-create-convert-sh.sh when pulling the conversion trigger for the first time. If, however, [global parameters](Reusing enhancement parameters for multiple versions or datasets) are present, manual/x.global.e1.params.ttl are [generated from the global parameters](Generating enhancement parameters) each time the conversion trigger is pulled. As the comments in manual/x.global.e1.params.ttl` suggest, they should NOT be edited by hand. The page about the retrieval phase of conversion process provides a good introduction to the directory structure conventions used and can help understand where [global parameters] need to be situated. When placed, the global parameter are recognized during Automated creation of a new Versioned Dataset.
FAQs elsewhere that should be consolidated to this page:
- http://logd.tw.rpi.edu/lab/faq/what_conventions_should_we_use_when_choosing_conversionsource_identifier_conversiondataset_identifier_and_conversionversion_identifier
- http://logd.tw.rpi.edu/lab/faq/how_do_we_identify_versions_abstract_dataset
- http://logd.tw.rpi.edu/lab/faq/why_csv2rdf4lods_cr-create-convert-shsh_not_creating_script_instead_just_printing_it_out