Replies: 1
Hi,
So I followed the horton works tutorial and completed the tutorial to set up Solr here: http://hortonworks.com/hadoop-tutorial/searching-data-solr/
This worked with no problems and I was able to index the CSV file which was suggested. All good.
Next I thought it would be good to index a directory full of random pdf files, excel files and csvs. I used the following command as reference at:
https://docs.lucidworks.com/display/help/Create+a+New+Hortonworks+Hadoop+Data+Source##
Command: hadoop jar /opt/solr/fusion/connectors/plugins/lucid.hadoop/hadoop-job/hadoop-lws-job-2.0.0-rc2-0.jar com.lucidworks.hadoop.ingest.IngestJob -Dlww.commit.on.close=true -cls com.lucidworks.hadoop.ingest.DirectoryIngestMapper -c hdp1 -i /user/solr/data/csv/ -of com.lucidworks.hadoop.io.LWMapRedOutputFormat -s http://localhost:8983/solr
The indexing failed and gave me an error as below:
File Input Format Counters
Bytes Read=149
File Output Format Counters
Bytes Written=0
Didn’t ingest any documents, failing
Can someone please help me out here? I’ve googled everywhere and cant find another instance of anyone using DirectoryIngestMapper