package src;
import java.util.Properties;
import edu.stanford.nlp.pipeline.StanfordCoreNLP;
public class NLPTest {
public static void main(String[] args){
Properties props = new Properties();
props.put("annotators", "tokenize, ssplit, pos, lemma, ner, parse, dcoref");
StanfordCoreNLP coreNLP = new StanfordCoreNLP(props);
}
}
I ran this sample code in my eclipse but it gives following error: Loading classifier from edu/stanford/nlp/models/ner/english.all.3class.distsim.crf.ser.gz ... Exception in thread "main" java.lang.OutOfMemoryError: Java heap space
Although everything works perfectly when I run Stanford CoreNLP from Command Propmt. Can anybody tell me the solution? Is it related to memory allocation to Eclipse?