I have a DB having text file attributes and text file primary key IDs and
indexed around 1 million text files along with their IDs (primary keys in DB).
Now, I am searching at two levels.
First is straight forward DB search, where i get primary keys as result (roughly 2 or 3 million IDs)
Then i make a Boolean query for instance as following
+Text:”test*” +(pkID:1 pkID:4 pkID:100 pkID:115 pkID:1041 …. )
and search it in my Index file.
The problem is that such query (having 2 million clauses) takes toooooo much time to give result and consumes reallly too much memory….
Is there any optimization solution for this problem ?
Assuming you can reuse the dbid part of your queries:
As long as the pkid search can be reused, you should quite a large improvement. As long as you don’t optimise your index, the effect of caching should even work through commit points (I understand the bit sets are calculated on a per-segment basis).
HTH
p.s.
I think it would be remiss of me not to note that I think you’re putting your index through all sorts of abuse by using it like this!