Java 为什么我的Android程序总是在内存中出错
大家好,我的应用程序涉及Lucene java库,我有一个问题,我不知道到底是什么错误 下面是错误控制台的示例 错误/AndroidRuntime(25909):java.lang.OutOfMemoryError:(堆大小=32775KB,分配=30112KB,位图大小=0KB) 错误/AndroidRuntime(25909):位于org.apache.lucene.index.FreqProxTermsWriterPerField$FreqProxPostingsArray。(FreqProxTermsWriterPerField.java:193) 错误/AndroidRuntime(25909):位于org.apache.lucene.index.FreqProxTermsWriterPerField$FreqProxPostingsArray.newInstance(FreqProxTermsWriterPerField.java:204) ERROR/AndroidRuntime(25909):位于org.apache.lucene.index.ParallelPostingsArray.grow(ParallelPostingsArray.java:48) ERROR/AndroidRuntime(25909):位于org.apache.lucene.index.TermsHashPerField.growParallelPostingsArray(TermsHashPerField.java:137) ERROR/AndroidRuntime(25909):位于org.apache.lucene.index.TermsHashPerField.add(TermsHashPerField.java:440) ERROR/AndroidRuntime(25909):位于org.apache.lucene.index.docinverterfield.processFields(docinverterfield.java:172) ERROR/AndroidRuntime(25909):位于org.apache.lucene.index.DocFieldProcessorPerThread.processDocument(DocFieldProcessorPerThread.java:278) 错误/AndroidRuntime(25909):位于org.apache.lucene.index.DocumentsWriter.updateDocument(DocumentsWriter.java:766) ERROR/AndroidRuntime(25909):位于org.apache.lucene.index.IndexWriter.addDocument(IndexWriter.java:2067) ERROR/AndroidRuntime(25909):位于org.apache.lucene.index.IndexWriter.addDocument(IndexWriter.java:2041) 它说的是indexwriter领域,我不知道它是什么 你们能帮帮我吗?谢谢你的答复 这是我的密码Java 为什么我的Android程序总是在内存中出错,java,lucene,Java,Lucene,大家好,我的应用程序涉及Lucene java库,我有一个问题,我不知道到底是什么错误 下面是错误控制台的示例 错误/AndroidRuntime(25909):java.lang.OutOfMemoryError:(堆大小=32775KB,分配=30112KB,位图大小=0KB) 错误/AndroidRuntime(25909):位于org.apache.lucene.index.FreqProxTermsWriterPerField$FreqProxPostingsArray。(FreqPr
public class CalculateWeightPage {
protected static Crawlers crawlers;
protected static StopWordsAndStemmer stemmer;
protected static CountWords countWords;
protected static StringSplitter splitter;
protected static ShortingStringArray shortingStringArray;
public static String[][] calulateRelevancePage(String[][] wkt,String urlPage) {
// 1.1.Defining parameters
int p = 0;
int count = 0;
int count2 = 0;
String title = "";
String body = "";
int titleFreq = 0;
int bodyFreq = 0;
String[][] wkp = null ;
int newTf = 0;
int y = 0;
int counter = 0;
try {
// 1.2.Extracting the text body and title from webPage
Map bodyTitle = crawlers.extractBodyAndTitle(urlPage);
if(bodyTitle.containsKey("title")){
title = stemmer.removeStopWordsAndStem(((String) bodyTitle.get("title")).toLowerCase());
body = stemmer.removeStopWordsAndStem(((String) bodyTitle.get("body")).toLowerCase());
// 1.4.Making a list containing unique words from text title and body
List bodyTitleUnique = splitter.StringUnique(body);
int sizeList = bodyTitleUnique.size();
wkp = new String[sizeList][2];
// 1.5.Calculating each tf
for (int r = 0; r < sizeList; r++) {
titleFreq = 0;
bodyFreq = 0;
// 1.5.1.Calculating tf in title
titleFreq = countWords.calculate(title, bodyTitleUnique.get(r).toString());
// 1.5.2.Calculating tf in body
bodyFreq = countWords.calculate(body, bodyTitleUnique.get(r).toString());
if (!(titleFreq == 0)) {
newTf = (titleFreq * 2) + (bodyFreq - titleFreq);
} else {
newTf = titleFreq + bodyFreq;
}
// 1.6.Inserting the result into string array
if(!(newTf == 0)){
wkp[r][0] = bodyTitleUnique.get(r).toString();
wkp[r][1] = String.valueOf(newTf);
}
}
}else{
return wkp;
}
} catch (Exception e) {
// TODO: handle exception
}
return wkp;
}
}您可能会遇到此错误,因为您需要的所有数据都无法放入内存中 然而,您的解决方案看起来有点过于复杂:您不需要Lucene来计算内存中的术语频率(CountWords的计算方法),您只需要分析输入并将频率存储在
HashMap
映射中
此外,尽管您的代码可能可以工作,但您的代码中有一些看起来不正确的地方:
- 您应该在优化之前调用commit,以便在优化之前创建段(尽管优化毫无意义,因为您将只有一个段)
- 您正在同一目录上打开一个索引搜索器和一个索引读取器,但打开索引搜索器会在引擎盖下打开一个索引读取器,因此实际上会打开两个等效的读取器,尽管只需要一个
- 关闭索引搜索器,但不关闭阅读器
public class CountWords {
CountWords() {
}
protected static StopWordsAndStemmer stemmer;
public static int calculate(String txt, String keyword) {
StopAnalyzer analyzer = new StopAnalyzer(Version.LUCENE_CURRENT);
RAMDirectory idx = new RAMDirectory();
int counts = 0;
int count = 0;
try {
IndexWriter writer = new IndexWriter(idx, analyzer, true,
IndexWriter.MaxFieldLength.UNLIMITED);
Document doc = new Document();
//String text1 = stemmer.removeStopWordsAndStem(txt.toLowerCase());
writer.addDocument(createDocument("", txt));
writer.optimize();
writer.close();
Searcher searcher = new IndexSearcher(idx);
IndexReader ir = IndexReader.open(idx);
TermDocs termDocs = ir.termDocs(new Term("content", keyword.toLowerCase()));
while (termDocs.next()) {
count = count + termDocs.freq();
}
//counts = count(count);
searcher.close();
} catch (IOException ioe) {
ioe.printStackTrace();
}
return count;
}
private static Document createDocument(String title, String content) {
Document doc = new Document();
doc.add(new Field("content", new StringReader(content)));
return doc;
}
private static int search(Searcher searcher, String queryString)throws ParseException, IOException {
StandardAnalyzer analyzer = new StandardAnalyzer(Version.LUCENE_CURRENT);
QueryParser parser = new QueryParser(Version.LUCENE_CURRENT, "content",analyzer);
Query query = parser.parse(queryString);
TopScoreDocCollector collector = TopScoreDocCollector.create(10, true);
searcher.search(query, collector);
return collector.getTotalHits();
}
public static Integer count(int count) {
if (count == 0) {
count = 1;
} else {
count = count;
}
return count;
}