Jost Baron | 31 Jul 01:29 2014
Picon
Picon

Re: Index a time/date range


Hi Ryan,

On 07/31/2014 01:26 AM, Ryan Cutter wrote:
> Is there a way to index time or date ranges?  That is, assume 2
> docs:
> 
> #1: date = 2014-01-01 #2: date = 2014-02-01 through 2014-05-01
> 
> Would there be a way to index #2's date as a single field and have
> all the search options you usually get with time/date?
> 
> One strategy could be to index the start and stop values
> separately.  Just wondering if there's a fancier option out there.

Take a look at this:

https://people.apache.org/~hossman/spatial-for-non-spatial-meetup-20130117/

Regards,
Jost

Alexandre Rafalovitch | 31 Jul 02:01 2014
Picon

Re: Index a time/date range

For fancier versions, some people used geo coordinates to represent start
on X axis and stop on Y. Then use perimeter bounds to do overlaps.

There was a discussion on the list about that a while ago.

Regards,
    Alex
On 31/07/2014 6:26 am, "Ryan Cutter" <ryancutter <at> gmail.com> wrote:

> Is there a way to index time or date ranges?  That is, assume 2 docs:
>
> #1: date = 2014-01-01
> #2: date = 2014-02-01 through 2014-05-01
>
> Would there be a way to index #2's date as a single field and have all the
> search options you usually get with time/date?
>
> One strategy could be to index the start and stop values separately.  Just
> wondering if there's a fancier option out there.
>
> Thanks, Ryan
>
Ryan Cutter | 31 Jul 01:26 2014
Picon

Index a time/date range

Is there a way to index time or date ranges?  That is, assume 2 docs:

#1: date = 2014-01-01
#2: date = 2014-02-01 through 2014-05-01

Would there be a way to index #2's date as a single field and have all the
search options you usually get with time/date?

One strategy could be to index the start and stop values separately.  Just
wondering if there's a fancier option out there.

Thanks, Ryan
Carlos Maroto | 31 Jul 01:07 2014
Picon

Setting a Key/Tag/Label for each group.query Result Set

Hi,

I'm trying to get results in a single Solr call through multiple
group.query definitions.  I'm getting the results I want but, each group is
presented under a "name" consisting of the query used for that group.

I'd like to change the "name" of each group to some meaningful name
instead.  I'm looking for something similar to the "key" feature in Facets
(
https://wiki.apache.org/solr/SimpleFacetParameters#key_:_Changing_the_output_key
)

For example, the current output I get is:
...
<lst name="grouped">
    <lst name="{!edismax qf=Keyword mm=100% v=$queryTerms}">
 <!-- Change this ... -->
        <int name="matches">5849</int>
        <result name="doclist" numFound="0" start="0"/>
    </lst>
    <lst name="{!edismax qf=synonyms mm='2<-1 5<100%' v=$queryTerms}"> <!--
... and also change this -->
    <int name="matches">5849</int>
        <result name="doclist" numFound="7" start="0">
            <doc>
...

Where I'd like to see something like:
...
<lst name="grouped">
(Continue reading)

Ameya Aware | 30 Jul 22:15 2014
Picon

Avoiding indexing of hidden folders and files

Hi,

I noticed a fact that Solr indexes all the folders and files including
hidden files.

Can anyone help me with avoiding indexing of hidden files?

Thanks,
Ameya
Ameya Aware | 30 Jul 20:09 2014
Picon

Exception : Processing of multipart/form-data request failed.

Hi

I am getting exception for  Processing of multipart/form-data request
failed.

My solrconfig.xml contains:

   <requestParsers enableRemoteStreaming="true"
                    multipartUploadLimitInKB="5120000"
                    formdataUploadLimitInKB="2048"
                    addHttpRequestToContext="false"/>

Please find below the stack trace.

ERROR - 2014-07-30 13:52:05.013; org.apache.solr.common.SolrException;
null:org.apache.commons.fileupload.FileUploadBase$IOFileUploadException:
Processing of multipart/form-data request failed. early EOF
at
org.apache.commons.fileupload.FileUploadBase.parseRequest(FileUploadBase.java:367)
at
org.apache.commons.fileupload.servlet.ServletFileUpload.parseRequest(ServletFileUpload.java:126)
at
org.apache.solr.servlet.SolrRequestParsers$MultipartRequestParser.parseParamsAndFillStreams(SolrRequestParsers.java:543)
at
org.apache.solr.servlet.SolrRequestParsers$StandardRequestParser.parseParamsAndFillStreams(SolrRequestParsers.java:679)
at
org.apache.solr.servlet.SolrRequestParsers.parse(SolrRequestParsers.java:146)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:398)
at
(Continue reading)

Eugene | 30 Jul 19:47 2014
Picon

Implementing custom analyzer for multi-language stemming

    Hello, fellow Solr and Lucene users and developers!

    In our project we receive text from users in different languages. We
detect language automatically and use Google Translate APIs a lot (so
having arbitrary number of languages in our system doesn't concern us).
However we need to be able to search using stemming. Having nearly hundred
of fields (several fields for each language with language-specific
stemmers) listed in our search query is not an option. So we need a way to
have a single index which has stemmed tokens for different languages. I
have two questions:

    1. Are there already (third-party) custom multi-language stemming
analyzers? (I doubt that no one else ran into this issue)

    2. If I'm going to implement such analyzer myself, could you please
suggest a better way to 'pass' detected language value into such analyzer?
Detecting language in analyzer itself is not an option, because: a) we
already detect it in other place b) we do it based on combined values of
many fields ('name', 'topic', 'description', etc.), while current field can
be to short for reliable detection c) sometimes we just want to specify
language explicitly. The obvious hack would be to prepend ISO 639-1 code to
field value. But I'd like to believe that Solr allows for cleaner solution.
I could think about either: a) custom query parameter (but I guess, it will
require modifying request handlers, etc. which is highly undesirable) b)
getting value from other field (we obviously have 'language' field and we
do not have mixed-language records). If it is possible, could you please
describe the mechanism for doing this or point to relevant code examples?
Thank you very much and have a good day!
Shawn Heisey | 30 Jul 15:27 2014

Index size increase after upgrade to 4.9?

Yesterday I upgraded my dev server to Solr 4.9, and also upgraded a
third-party plugin to a new version that's compatible with Solr 4.9.

After the index was rebuilt, each shard was 28GB ... but before the
upgrade, each shard was only 20GB.  The number of documents per shard
(16.4 million) actually went *down* a little bit, and the config/schema
hasn't changed.

Could this be explained by the new Solr version?  I've also asked the
third-party plugin company about this problem.

Thanks,
Shawn

Liram Vardi | 30 Jul 15:25 2014
Picon

Identify specific document insert error inside a solrj batch request

Hi All,

I have a question regarding the use of HttpSolrServer (SolrJ).
I have a collection of SolrInputDocuments I want to send to Solr as a batch.
Now, let's assume that one of the docs inside this collection is corrupted (missing some "required" field).
When I send the batch of docs to solr using HttpSolrServer.add(Collection< SolrInputDocument> docs) I am
getting the following general exception:

"org.apache.solr.client.solrj.impl.HttpSolrServer$RemoteSolrException:
Server at http://172.23.3.91:8210/solr/template returned non ok status:500, message:Server Error"

When I check Solr log, I can identify exactly which is the corrupted document.

My question:
Is it possible to identify the problematic document at the client side? (for recovery purposes)

Thanks,
Liram
Pbbhoge | 30 Jul 14:59 2014
Picon

Search on Date Field

In my SOLR there is date field(published_date) and values are in this format
"2012-09-26T10:08:09.123Z"

How I can search by simple input like "2012-09-10" instead of full ISO date
format.

Is it possible in SOLR?

--
View this message in context: http://lucene.472066.n3.nabble.com/Search-on-Date-Field-tp4150076.html
Sent from the Solr - User mailing list archive at Nabble.com.

Tommaso Teofili | 30 Jul 15:00 2014
Picon

Tika analyzers

Hi all,

while SolrCell works nicely when in need of indexing binary documents, I am
wondering about the possibility of having Lucene / Solr documents that have
binaries in specific Lucene fields, e.g. title="a nice doc",
name"blabla.doc", binary="0x1234...".

In that case the "binary" field should have an indexing analyzer which can
extract the text from the binary and index it.

Would it make sense to create a Tika based analyzer for that purpose?

Regards,
Tommaso

Gmane