1 |
| |
2 |
| |
3 |
| |
4 |
| |
5 |
| |
6 |
| |
7 |
| |
8 |
| |
9 |
| |
10 |
| |
11 |
| |
12 |
| |
13 |
| |
14 |
| |
15 |
| |
16 |
| |
17 |
| |
18 |
| |
19 |
| |
20 |
| package org.apache.xindice.core.query.ftsearch; |
21 |
| |
22 |
| import org.apache.lucene.index.IndexReader; |
23 |
| import org.apache.lucene.index.TermFreqVector; |
24 |
| import org.apache.lucene.index.TermEnum; |
25 |
| import org.apache.lucene.index.Term; |
26 |
| import org.apache.lucene.index.TermDocs; |
27 |
| import org.apache.lucene.index.TermPositions; |
28 |
| import org.apache.lucene.index.TermVectorMapper; |
29 |
| import org.apache.lucene.document.Document; |
30 |
| import org.apache.lucene.document.FieldSelector; |
31 |
| import org.apache.lucene.analysis.Analyzer; |
32 |
| import org.apache.lucene.analysis.TokenStream; |
33 |
| import org.apache.lucene.analysis.tokenattributes.TermAttribute; |
34 |
| import org.apache.lucene.search.DefaultSimilarity; |
35 |
| import org.apache.xindice.xml.dom.NodeImpl; |
36 |
| import org.w3c.dom.NodeList; |
37 |
| import org.w3c.dom.Node; |
38 |
| |
39 |
| import java.io.IOException; |
40 |
| import java.io.StringReader; |
41 |
| import java.util.Collection; |
42 |
| import java.util.Arrays; |
43 |
| import java.util.ArrayList; |
44 |
| import java.util.SortedMap; |
45 |
| import java.util.TreeMap; |
46 |
| import java.util.HashSet; |
47 |
| import java.util.List; |
48 |
| |
49 |
| |
50 |
| |
51 |
| |
52 |
| |
53 |
| |
54 |
| |
55 |
| public class NodeReader extends IndexReader { |
56 |
| |
57 |
| private final ArrayList[] nodes; |
58 |
| |
59 |
| private byte[] norms; |
60 |
| |
61 |
| |
62 |
| private final SortedMap termMap; |
63 |
| |
64 |
54
| protected NodeReader(NodeList list, Analyzer analyzer) {
|
65 |
54
| super();
|
66 |
| |
67 |
54
| nodes = new ArrayList[list.getLength()];
|
68 |
| |
69 |
54
| for (int i = 0; i < nodes.length; i++) {
|
70 |
74
| nodes[i] = new ArrayList();
|
71 |
74
| Node node = list.item(i);
|
72 |
| |
73 |
74
| String text;
|
74 |
74
| if (node instanceof NodeImpl) {
|
75 |
| |
76 |
74
| text = ((NodeImpl) node).getTextContent();
|
77 |
| } else { |
78 |
0
| text = getTextContent(node);
|
79 |
| } |
80 |
| |
81 |
74
| TokenStream stream = analyzer.tokenStream("", new StringReader(text));
|
82 |
74
| TermAttribute termAttr = (TermAttribute) stream.addAttribute(TermAttribute.class);
|
83 |
74
| try {
|
84 |
74
| stream.reset();
|
85 |
74
| while (stream.incrementToken()) {
|
86 |
1256
| nodes[i].add(termAttr.term());
|
87 |
| } |
88 |
74
| stream.end();
|
89 |
74
| stream.close();
|
90 |
| } catch (IOException e) { |
91 |
| |
92 |
| } |
93 |
| } |
94 |
| |
95 |
| |
96 |
54
| norms = new byte[nodes.length];
|
97 |
54
| Arrays.fill(norms, DefaultSimilarity.encodeNorm(1.0f));
|
98 |
| |
99 |
| |
100 |
54
| termMap = buildTermMap();
|
101 |
| } |
102 |
| |
103 |
0
| public TermFreqVector[] getTermFreqVectors(int docNumber) {
|
104 |
0
| throw new UnsupportedOperationException();
|
105 |
| } |
106 |
| |
107 |
0
| public TermFreqVector getTermFreqVector(int docNumber, String field) {
|
108 |
0
| throw new UnsupportedOperationException();
|
109 |
| } |
110 |
| |
111 |
0
| public void getTermFreqVector(int i, String string, TermVectorMapper termVectorMapper) throws IOException {
|
112 |
0
| throw new UnsupportedOperationException();
|
113 |
| } |
114 |
| |
115 |
0
| public void getTermFreqVector(int i, TermVectorMapper termVectorMapper) throws IOException {
|
116 |
0
| throw new UnsupportedOperationException();
|
117 |
| } |
118 |
| |
119 |
54
| public int numDocs() {
|
120 |
54
| return nodes.length;
|
121 |
| } |
122 |
| |
123 |
116
| public int maxDoc() {
|
124 |
116
| return nodes.length;
|
125 |
| } |
126 |
| |
127 |
| |
128 |
| |
129 |
| |
130 |
0
| public Document document(int n, FieldSelector fieldSelector) {
|
131 |
0
| return null;
|
132 |
| } |
133 |
| |
134 |
| |
135 |
| |
136 |
| |
137 |
0
| public boolean isDeleted(int n) {
|
138 |
0
| return false;
|
139 |
| } |
140 |
| |
141 |
| |
142 |
| |
143 |
| |
144 |
0
| public boolean hasDeletions() {
|
145 |
0
| return false;
|
146 |
| } |
147 |
| |
148 |
53
| public byte[] norms(String field) throws IOException {
|
149 |
53
| return field.length() == 0 ? norms : null;
|
150 |
| } |
151 |
| |
152 |
0
| public void norms(String field, byte[] bytes, int offset) {
|
153 |
0
| System.arraycopy(norms, 0, bytes, offset, maxDoc());
|
154 |
| } |
155 |
| |
156 |
0
| protected void doSetNorm(int doc, String field, byte value) {
|
157 |
0
| if (field.length() > 0) {
|
158 |
0
| return;
|
159 |
| } |
160 |
| |
161 |
0
| norms[doc] = value;
|
162 |
| } |
163 |
| |
164 |
0
| public TermEnum terms() {
|
165 |
0
| return new NodeTermEnum(termMap);
|
166 |
| } |
167 |
| |
168 |
11
| public TermEnum terms(Term t) {
|
169 |
11
| return new NodeTermEnum(termMap, t);
|
170 |
| } |
171 |
| |
172 |
| |
173 |
| |
174 |
| |
175 |
| |
176 |
| |
177 |
54
| private SortedMap buildTermMap() {
|
178 |
54
| SortedMap map = new TreeMap();
|
179 |
| |
180 |
54
| for (int i = 0; i < nodes.length; i++) {
|
181 |
74
| for (int j = 0; j < nodes[i].size(); j++) {
|
182 |
1256
| String term = (String) nodes[i].get(j);
|
183 |
| |
184 |
1256
| List docs;
|
185 |
1256
| if (map.containsKey(term)) {
|
186 |
522
| docs = (List) map.get(term);
|
187 |
| } else { |
188 |
734
| docs = new ArrayList();
|
189 |
| } |
190 |
| |
191 |
1256
| docs.add(new Integer(i));
|
192 |
1256
| map.put(term, docs);
|
193 |
| } |
194 |
| } |
195 |
| |
196 |
54
| return map;
|
197 |
| } |
198 |
| |
199 |
59
| public int docFreq(Term t) {
|
200 |
59
| List docs = (List) termMap.get(t.text());
|
201 |
| |
202 |
| |
203 |
59
| if (docs == null) {
|
204 |
22
| return 0;
|
205 |
| } |
206 |
| |
207 |
37
| HashSet set = new HashSet();
|
208 |
37
| set.addAll(docs);
|
209 |
37
| return set.size();
|
210 |
| } |
211 |
| |
212 |
55
| public TermDocs termDocs() {
|
213 |
55
| return new NodeTermDocs(this);
|
214 |
| } |
215 |
| |
216 |
4
| public TermPositions termPositions() {
|
217 |
4
| return new NodeTermPositions(this);
|
218 |
| } |
219 |
| |
220 |
| |
221 |
| |
222 |
| |
223 |
0
| protected void doDelete(int docNum) {
|
224 |
0
| throw new UnsupportedOperationException();
|
225 |
| } |
226 |
| |
227 |
| |
228 |
| |
229 |
| |
230 |
0
| protected void doUndeleteAll() {
|
231 |
0
| throw new UnsupportedOperationException();
|
232 |
| } |
233 |
| |
234 |
| |
235 |
| |
236 |
| |
237 |
0
| protected void doCommit() {
|
238 |
| } |
239 |
| |
240 |
| |
241 |
| |
242 |
| |
243 |
0
| protected void doClose() {
|
244 |
| } |
245 |
| |
246 |
| |
247 |
| |
248 |
| |
249 |
0
| public Collection getFieldNames(FieldOption fldOption) {
|
250 |
0
| throw new UnsupportedOperationException();
|
251 |
| } |
252 |
| |
253 |
59
| ArrayList[] getNodes() {
|
254 |
59
| return nodes;
|
255 |
| } |
256 |
| |
257 |
59
| SortedMap getTermMap() {
|
258 |
59
| return termMap;
|
259 |
| } |
260 |
| |
261 |
| |
262 |
| |
263 |
| |
264 |
| |
265 |
| |
266 |
| |
267 |
0
| private String getTextContent(Node node) {
|
268 |
0
| String text = null;
|
269 |
0
| switch (node.getNodeType()) {
|
270 |
0
| case Node.ATTRIBUTE_NODE:
|
271 |
0
| case Node.CDATA_SECTION_NODE:
|
272 |
0
| case Node.COMMENT_NODE:
|
273 |
0
| case Node.PROCESSING_INSTRUCTION_NODE:
|
274 |
0
| case Node.TEXT_NODE:
|
275 |
0
| text = node.getNodeValue();
|
276 |
0
| break;
|
277 |
0
| case Node.ELEMENT_NODE:
|
278 |
0
| case Node.DOCUMENT_FRAGMENT_NODE:
|
279 |
0
| case Node.ENTITY_NODE:
|
280 |
0
| case Node.ENTITY_REFERENCE_NODE:
|
281 |
0
| StringBuffer val = new StringBuffer();
|
282 |
| |
283 |
0
| NodeList children = node.getChildNodes();
|
284 |
0
| if (children == null || children.getLength() == 0) {
|
285 |
0
| text = "";
|
286 |
0
| break;
|
287 |
| } |
288 |
| |
289 |
0
| for (int i = 0; i < children.getLength(); i++) {
|
290 |
0
| val.append(getTextContent(children.item(i)));
|
291 |
| } |
292 |
0
| text = val.toString();
|
293 |
0
| break;
|
294 |
| } |
295 |
| |
296 |
0
| return text;
|
297 |
| } |
298 |
| } |