Can not find schema.xml file for solrcloud - schema

I want to change schema.xml file, but i can not find it. How to change schema.xml file?
link image : https://drive.google.com/file/d/0BxOsysEbbmMOQ0hoV19KaVo0SUU/view?usp=sharing

As per the image, you are working with the SolrCloud example Collection gettingstarted.
In SolrCloud schema.xml file is maintained by ZooKeeper and so when you change the schema, you will have to upload to ZooKeeper to make it effective.
Please describe in detail what you are looking for?
Note: I still don't have the privilege of comments, or else this should be in comments.

Related

Apache ManifoldCF TIKA

I am trying to extract the text content of a PDF using the Apache Tika integration on Apache ManifoldCF, in order to ingest some PDF files on my Laptop in an Elasticsearch server.
After properly creating the Tika Transformer and configuring it inside my job, I see that the resulting field "_content" on ES is filled with the binary encoding of the file, and not the text.
I saw also this :Extract file content with ManifoldCF, But still no answer has been provided (since 2015!).
Can anybody help me?
Thanks!
In the output connector for elastic search what is the field name that you have specified for the content field?
Please provide a field name as well as max document size.

What is the significance of data-config.xml file in Solr?

and when shall I use it? How is it configured can anyone please tell me in detail?
The data-config.xml file is an example configuration file for how to use the DataImportHandler in Solr. It's one way of getting data into Solr, allowing one of the servers to connect through JDBC (or through a few other plugins) to a database server or a set of files and import them into Solr.
DIH has a few issues (for example the non-distributed way it works), so it's usually suggested to write the indexing code yourself (and POST it to Solr from a suitable client, such as SolrJ, Solarium, SolrClient, MySolr, etc.)
It has been mentioned that the DIH functionality really should be moved into a separate application, but that hasn't happened yet as far as I know.

Apache Solr & schema.xml

I have just began with Apache Solr 4.1 yesterday and though I have managed to get our MySQL data imported successfully in Solr, I am unable to view any data using queries. I suspect the problem is in schema.xml changes (data-config.xml is correct). Here are my questions -
Do I need to add all DB fields in the schema.xml? My table has 275+
fields, and configuring all of them would be a task. I am hoping
there is a way to auto-configure these fields.
Is there a way to use separate schema.xml for my requirement? Where
and how do I configure this? I don't want to modify the example-DIH's
sample schema.xml
Any pointers here would be highly appreciated! I have already gone through this document -
http://wiki.apache.org/solr/DataImportHandler
and I have also read few question posted here, but couldn't find answer to my queries.
1) All the fields that you are mentioning in data-config.xml as field column must add them to schema.xml. No auto-configure is available.
2) No we cannot use separate schema.xml.
Could you please show some of your indexed field that you have mentioned in schema.xml

Can't get ApacheDS to load the example data file

In ApacheDS, I'm trying to load the sample LDAP data provided in ldif format and with instructions here.
I followed the same pattern I used to load the context record by specifying an ldif file in server.xml for the partition (directions, albiet lousy ones, located here).
So I have...
<apacheDS id="apacheDS">
<ldapServer>#ldapServer</ldapServer>
<ldifDirectory>sevenSeasRoot.ldif</ldifDirectory>
<ldifDirectory>apache_ds_tutorial.ldif</ldifDirectory>
</apacheDS>
The sevenSeasRoot.ldif file seems to have loaded properly, because I can see an entry for it in LdapBrowser. But there are no records under it.
What am I doing wrong? How do I configure server.xml to load the child records for sevenSeas?
Just a quick note, the config says ldif*Directory* but you are passing a file.
OTOH, I guess you are using 1.5.7 this is very old, it would be better to try the latest version for better support.

Lucene and Solr

I am creating the Index using Lucene.
But using the SolrSearch engine to search.
My problem is
while I index I add each filed in my
code using the code
doc.add(fieldname, val, tokenized)
**But my code does not see the schema file
Even copy fields I need to add manually**
Now I want to use the autosuggest feature of Solr
I do not know how to enable this feature while creating the Index
But when I use the simplePostTool to post the data through Solr all is fine.
I cannot do that because I have to
get some text from different urls.
So can someone please advise me how
can I achieve this? A sample code
will be very helpful. In any case If
I can have some code that can see
the schema file and use the
fieldTypes it would be great.
Thanks everyone.
--pramila
See EmbeddedSolrServer at:
http://wiki.apache.org/solr/Solrj
It's a pure Java API to Solr, which will allow you to index your documents using the schema.xml you have defined.