Quantcast
Channel: Hortonworks » All Topics
Viewing all articles
Browse latest Browse all 5121

Solr, Lucidworks & indexing files

$
0
0

Replies: 1

Hi,

So I followed the horton works tutorial and completed the tutorial to set up Solr here: http://hortonworks.com/hadoop-tutorial/searching-data-solr/
This worked with no problems and I was able to index the CSV file which was suggested. All good.
Next I thought it would be good to index a directory full of random pdf files, excel files and csvs. I used the following command as reference at:

https://docs.lucidworks.com/display/help/Create+a+New+Hortonworks+Hadoop+Data+Source##

Command: hadoop jar /opt/solr/fusion/connectors/plugins/lucid.hadoop/hadoop-job/hadoop-lws-job-2.0.0-rc2-0.jar com.lucidworks.hadoop.ingest.IngestJob -Dlww.commit.on.close=true -cls com.lucidworks.hadoop.ingest.DirectoryIngestMapper -c hdp1 -i /user/solr/data/csv/ -of com.lucidworks.hadoop.io.LWMapRedOutputFormat -s http://localhost:8983/solr

The indexing failed and gave me an error as below:

File Input Format Counters
Bytes Read=149
File Output Format Counters
Bytes Written=0
Didn’t ingest any documents, failing

Can someone please help me out here? I’ve googled everywhere and cant find another instance of anyone using DirectoryIngestMapper :-(


Viewing all articles
Browse latest Browse all 5121

Trending Articles



<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>