| Chunk |
|---|
| Conflicting content |
|---|
import org.apache.lucene.analysis.Analyzer; import org.apache.lucene.analysis.TokenStream; import org.apache.lucene.analysis.tokenattributes.CharTermAttribute; <<<<<<< HEAD ======= import org.apache.mahout.classifier.ClassifierResult; >>>>>>> 678d92be45331bdd660b941831296676aa1233e6 import org.apache.solr.common.SolrInputDocument; import org.apache.solr.common.SolrInputField; import org.apache.solr.update.AddUpdateCommand; |
| Solution content |
|---|
import org.apache.lucene.analysis.Analyzer; import org.apache.lucene.analysis.TokenStream; import org.apache.lucene.analysis.tokenattributes.CharTermAttribute; import org.apache.solr.common.SolrInputDocument; import org.apache.solr.common.SolrInputField; import org.apache.solr.update.AddUpdateCommand; |
| File |
|---|
| BayesUpdateRequestProcessor.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| Import |
| Chunk |
|---|
| Conflicting content |
|---|
package com.tamingtext.classifier.bayes; <<<<<<< HEAD import java.io.BufferedReader; import java.io.File; import java.io.IOException; import java.io.InputStreamReader; import java.util.Arrays; import java.util.Collection; import java.util.Collections; import java.util.HashMap; import java.util.LinkedList; import java.util.List; import java.util.Map; ======= >>>>>>> 678d92be45331bdd660b941831296676aa1233e6 import org.apache.commons.cli2.CommandLine; import org.apache.commons.cli2.Group; import org.apache.commons.cli2.Option; |
| Solution content |
|---|
package com.tamingtext.classifier.bayes; import java.io.BufferedReader; import java.io.File; import java.io.IOException; import java.io.InputStreamReader; import java.util.Arrays; import java.util.Collection; import java.util.Collections; import java.util.HashMap; import java.util.LinkedList; import java.util.List; import java.util.Map; import org.apache.commons.cli2.CommandLine; import org.apache.commons.cli2.Group; import org.apache.commons.cli2.Option; |
| File |
|---|
| ExtractTrainingData.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| Import |
| Chunk |
|---|
| Conflicting content |
|---|
import org.apache.hadoop.util.Tool; import org.apache.hadoop.util.ToolRunner; import org.apache.lucene.document.Document; <<<<<<< HEAD import org.apache.lucene.document.Fieldable; ======= import org.apache.lucene.index.DirectoryReader; import org.apache.lucene.index.DocsEnum; >>>>>>> 678d92be45331bdd660b941831296676aa1233e6 import org.apache.lucene.index.IndexReader; import org.apache.lucene.index.IndexableField; import org.apache.lucene.index.MultiFields; |
| Solution content |
|---|
import org.apache.hadoop.util.ToolRunner; import org.apache.hadoop.util.Tool; import org.apache.lucene.document.Document; import org.apache.lucene.document.Fieldable; import org.apache.lucene.index.IndexReader; |
| File |
|---|
| ExtractTrainingData.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| Import |
| Chunk |
|---|
| Conflicting content |
|---|
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
<<<<<<< HEAD
import com.google.common.io.Closeables;
/** A utility to extract training data from a Lucene index using document term vectors to recreate the list of terms
* found in each document. Writes output in Mahout Document Sequence File Format with the category and id as the key and the text as the value */
public class ExtractTrainingData extends Configured implements Tool {
private static final Logger log = LoggerFactory.getLogger(ExtractTrainingData.class);
private final Map |
| Solution content |
|---|
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import com.google.common.io.Closeables;
/** A utility to extract training data from a Lucene index using document term vectors to recreate the list of terms
* found in each document. Writes output in Mahout Document Sequence File Format with the category and id as the key and the text as the value */
public class ExtractTrainingData extends Configured implements Tool {
private static final Logger log = LoggerFactory.getLogger(ExtractTrainingData.class);
private final Map |
| File |
|---|
| ExtractTrainingData.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| Attribute |
| Class signature |
| Comment |
| Import |
| Method declaration |
| Method invocation |
| Method signature |
| Variable |
| Chunk |
|---|
| Conflicting content |
|---|
GroupBuilder gbuilder = new GroupBuilder();
Option inputOpt = obuilder.withLongName("dir")
<<<<<<< HEAD
.withRequired(true)
.withArgument(
abuilder.withName("dir")
.withMinimum(1)
.withMaximum(1).create())
=======
.withRequired(true)
.withArgument(
abuilder.withName("dir")
.withMinimum(1)
.withMaximum(1).create())
>>>>>>> 678d92be45331bdd660b941831296676aa1233e6
.withDescription("Lucene index directory containing input data")
.withShortName("d").create();
|
| Solution content |
|---|
GroupBuilder gbuilder = new GroupBuilder();
Option inputOpt = obuilder.withLongName("dir")
.withRequired(true)
.withArgument(
abuilder.withName("dir")
.withMinimum(1)
.withMaximum(1).create())
.withDescription("Lucene index directory containing input data")
.withShortName("d").create();
|
| File |
|---|
| ExtractTrainingData.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| Method invocation |
| Chunk |
|---|
| Conflicting content |
|---|
.withShortName("d").create();
Option categoryOpt = obuilder.withLongName("categories")
<<<<<<< HEAD
.withRequired(true)
.withArgument(
abuilder.withName("file")
.withMinimum(1)
.withMaximum(1).create())
=======
.withRequired(true)
.withArgument(
abuilder.withName("file")
.withMinimum(1)
.withMaximum(1).create())
>>>>>>> 678d92be45331bdd660b941831296676aa1233e6
.withDescription("File containing a list of categories")
.withShortName("c").create();
|
| Solution content |
|---|
.withShortName("d").create();
Option categoryOpt = obuilder.withLongName("categories")
.withRequired(true)
.withArgument(
abuilder.withName("file")
.withMinimum(1)
.withMaximum(1).create())
.withDescription("File containing a list of categories")
.withShortName("c").create();
|
| File |
|---|
| ExtractTrainingData.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| Method invocation |
| Chunk |
|---|
| Conflicting content |
|---|
.create();
=======
.withShortName("c").create();
Option outputOpt = obuilder.withLongName("output")
<<<<<<< HEAD
.withRequired(false)
.withArgument(
abuilder.withName("output")
.withMinimum(1)
.withMaximum(1).create())
.withDescription("Output directory")
.withShortName("o").create();
Option categoryFieldsOpt =
obuilder.withLongName("category-fields")
.withRequired(true)
.withArgument(
abuilder.withName("fields")
.withMinimum(1)
.withMaximum(1)
.create())
.withDescription("Fields to match categories against (comma-delimited)")
.withShortName("cf").create();
Option textFieldsOpt =
obuilder.withLongName("text-fields")
.withRequired(true)
.withArgument(
abuilder.withName("fields")
.withMinimum(1)
.withMaximum(1)
.create())
.withDescription("Fields from which to extract training text (comma-delimited)")
.withShortName("tf").create();
Option useTermVectorsOpt = obuilder.withLongName("use-term-vectors")
.withDescription("Extract term vectors containing preprocessed data " +
"instead of unprocessed, stored text values")
.withShortName("tv").create();
Option helpOpt = obuilder.withLongName("help")
.withDescription("Print out help")
.withShortName("h").create();
Group group = gbuilder.withName("Options")
.withOption(inputOpt)
.withOption(categoryOpt)
.withOption(outputOpt)
.withOption(categoryFieldsOpt)
.withOption(textFieldsOpt)
.withOption(useTermVectorsOpt)
.withRequired(false)
.withArgument(
abuilder.withName("output")
.withMinimum(1)
.withMaximum(1).create())
.withDescription("Output directory")
.withShortName("o").create();
Option categoryFieldsOpt =
obuilder.withLongName("category-fields")
.withRequired(true)
.withArgument(
abuilder.withName("fields")
.withMinimum(1)
.withMaximum(1)
.create())
.withDescription("Fields to match categories against (comma-delimited)")
.withShortName("cf").create();
Option textFieldsOpt =
obuilder.withLongName("text-fields")
.withRequired(true)
.withArgument(
abuilder.withName("fields")
.withMinimum(1)
.withMaximum(1)
.create())
.withDescription("Fields from which to extract training text (comma-delimited)")
.withShortName("tf").create();
Option useTermVectorsOpt = obuilder.withLongName("use-term-vectors")
.withDescription("Extract term vectors containing preprocessed data " +
"instead of unprocessed, stored text values")
.withShortName("tv").create();
Option helpOpt = obuilder.withLongName("help")
.withDescription("Print out help")
.withShortName("h").create();
Group group = gbuilder.withName("Options")
.withOption(inputOpt)
.withOption(categoryOpt)
.withOption(outputOpt)
.withOption(categoryFieldsOpt)
.withOption(textFieldsOpt)
.withOption(useTermVectorsOpt)
.create();
>>>>>>> 678d92be45331bdd660b941831296676aa1233e6
try {
Parser parser = new Parser(); |
| Solution content |
|---|
.withShortName("c").create();
Option outputOpt = obuilder.withLongName("output")
.withRequired(false)
.withArgument(
abuilder.withName("output")
.withMinimum(1)
.withMaximum(1).create())
.withDescription("Output directory")
.withShortName("o").create();
Option categoryFieldsOpt =
obuilder.withLongName("category-fields")
.withRequired(true)
.withArgument(
abuilder.withName("fields")
.withMinimum(1)
.withMaximum(1)
.create())
.withDescription("Fields to match categories against (comma-delimited)")
.withShortName("cf").create();
Option textFieldsOpt =
obuilder.withLongName("text-fields")
.withRequired(true)
.withArgument(
abuilder.withName("fields")
.withMinimum(1)
.withMaximum(1)
.create())
.withDescription("Fields from which to extract training text (comma-delimited)")
.withShortName("tf").create();
Option useTermVectorsOpt = obuilder.withLongName("use-term-vectors")
.withDescription("Extract term vectors containing preprocessed data " +
"instead of unprocessed, stored text values")
.withShortName("tv").create();
Option helpOpt = obuilder.withLongName("help")
.withDescription("Print out help")
.withShortName("h").create();
Group group = gbuilder.withName("Options")
.withOption(inputOpt)
.withOption(categoryOpt)
.withOption(outputOpt)
.withOption(categoryFieldsOpt)
.withOption(textFieldsOpt)
.withOption(useTermVectorsOpt)
.create();
try {
Parser parser = new Parser(); |
| File |
|---|
| ExtractTrainingData.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| Method invocation |
| Variable |
| Chunk |
|---|
| Conflicting content |
|---|
return 1;
}
<<<<<<< HEAD
Path inputDir = new Path(cmdLine.getValue(inputOpt).toString());
if (!fs.getFileStatus(inputDir).isDir()) {
throw new IllegalArgumentException(inputDir + " does not exist or is not a directory");
}
Path categoryFile = new Path(cmdLine.getValue(categoryOpt).toString());
if (!fs.isFile(categoryFile)) {
throw new IllegalArgumentException(categoryFile + " does not exist or is not a directory");
}
Path outputDir = new Path(cmdLine.getValue(outputOpt).toString());
fs.mkdirs(outputDir);
if (!fs.getFileStatus(outputDir).isDir()) {
throw new IllegalArgumentException("the output directory for " + outputDir + " is not a directory or could not be created");
=======
File inputDir = new File(cmdLine.getValue(inputOpt).toString());
if (!inputDir.isDirectory()) {
throw new IllegalArgumentException(inputDir + " does not exist or is not a directory");
}
File categoryFile = new File(cmdLine.getValue(categoryOpt).toString());
if (!categoryFile.isFile()) {
throw new IllegalArgumentException(categoryFile + " does not exist or is not a directory");
}
File outputDir = new File(cmdLine.getValue(outputOpt).toString());
outputDir.mkdirs();
if (!outputDir.isDirectory()) {
throw new IllegalArgumentException(outputDir + " is not a directory or could not be created");
>>>>>>> 678d92be45331bdd660b941831296676aa1233e6
}
Collection |
| Solution content |
|---|
return 1;
}
Path inputDir = new Path(cmdLine.getValue(inputOpt).toString());
if (!fs.getFileStatus(inputDir).isDir()) {
throw new IllegalArgumentException(inputDir + " does not exist or is not a directory");
}
Path categoryFile = new Path(cmdLine.getValue(categoryOpt).toString());
if (!fs.isFile(categoryFile)) {
throw new IllegalArgumentException(categoryFile + " does not exist or is not a directory");
}
Path outputDir = new Path(cmdLine.getValue(outputOpt).toString());
fs.mkdirs(outputDir);
if (!fs.getFileStatus(outputDir).isDir()) {
throw new IllegalArgumentException("the output directory for " + outputDir + " is not a directory or could not be created");
}
Collection |
| File |
|---|
| ExtractTrainingData.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| If statement |
| Method invocation |
| Throw statement |
| Variable |
| Chunk |
|---|
| Conflicting content |
|---|
boolean useTermVectors = cmdLine.hasOption(useTermVectorsOpt);
<<<<<<< HEAD
extractTraininingData(fs, conf, inputDir, categoryFile, categoryFields, textFields, outputDir, useTermVectors);
=======
extractTraininingData(inputDir, categoryFile, categoryFields, textFields, outputDir, useTermVectors);
>>>>>>> 678d92be45331bdd660b941831296676aa1233e6
} catch (OptionException e) {
log.error("Exception", e); |
| Solution content |
|---|
boolean useTermVectors = cmdLine.hasOption(useTermVectorsOpt);
extractTraininingData(fs, conf, inputDir, categoryFile, categoryFields, textFields, outputDir, useTermVectors);
} catch (OptionException e) {
log.error("Exception", e); |
| File |
|---|
| ExtractTrainingData.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| Method invocation |
| Chunk |
|---|
| Conflicting content |
|---|
* @param outputDir directory to write output to
* @throws IOException
*/
<<<<<<< HEAD
public void extractTraininingData(FileSystem fs, Configuration conf, Path indexDir, Path categoryFile,
Collection |
| Solution content |
|---|
* @throws IOException
*/
public void extractTraininingData(FileSystem fs, Configuration conf, Path indexDir, Path categoryFile,
Collection |
| File |
|---|
| ExtractTrainingData.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| Method signature |
| Chunk |
|---|
| Conflicting content |
|---|
log.info("Text fields: " + textFields.toString());
log.info("Use Term Vectors?: " + useTermVectors);
OpenObjectIntHashMap |
| Solution content |
|---|
log.info("Text fields: " + textFields.toString());
log.info("Use Term Vectors?: " + useTermVectors);
OpenObjectIntHashMap |
| File |
|---|
| ExtractTrainingData.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| Comment |
| For statement |
| If statement |
| Method invocation |
| Variable |
| Chunk |
|---|
| Conflicting content |
|---|
// determine whether any of the fields in this document contain a
// category in the category list
<<<<<<< HEAD
fields: for (String field: categoryFields) {
for (Fieldable f: d.getFieldables(field)) {
if (f.isStored() && !f.isBinary()) {
=======
fields:
for (String field : categoryFields) {
for (IndexableField f : d.getFields(field)) {
if (f.stringValue() != null) {
>>>>>>> 678d92be45331bdd660b941831296676aa1233e6
String fieldValue = f.stringValue().toLowerCase();
for (String cat : categories.keySet()) {
List |
| Solution content |
|---|
// determine whether any of the fields in this document contain a
// category in the category list
fields: for (String field: categoryFields) {
for (Fieldable f: d.getFieldables(field)) {
if (f.isStored() && !f.isBinary()) {
String fieldValue = f.stringValue().toLowerCase();
for (String cat: categories.keySet()) {
List |
| File |
|---|
| ExtractTrainingData.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| For statement |
| If statement |
| Chunk |
|---|
| Conflicting content |
|---|
buf.setLength(0);
for (String field : textFields) {
if (useTermVectors) {
<<<<<<< HEAD
appendVectorTerms(buf, reader.getTermFreqVector(i, field));
}
else {
appendFieldText(buf, d.getFieldable(field));
=======
appendVectorTerms(buf, reader.getTermVector(i, field), liveDocs);
} else {
appendFieldText(buf, d.getField(field));
>>>>>>> 678d92be45331bdd660b941831296676aa1233e6
}
}
key.set(i + "/" + category); |
| Solution content |
|---|
buf.setLength(0);
for (String field: textFields) {
if (useTermVectors) {
appendVectorTerms(buf, reader.getTermFreqVector(i, field));
}
else {
appendFieldText(buf, d.getFieldable(field));
}
}
key.set(i + "/" + category); |
| File |
|---|
| ExtractTrainingData.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| Method invocation |
| Chunk |
|---|
| Conflicting content |
|---|
closeWriters();
}
}
<<<<<<< HEAD
=======
>>>>>>> 678d92be45331bdd660b941831296676aa1233e6
if (log.isInfoEnabled()) {
StringBuilder b = new StringBuilder();
b.append("\nCatagory document counts:\n"); |
| Solution content |
|---|
}
}
closeWriters();
if (log.isInfoEnabled()) {
StringBuilder b = new StringBuilder();
b.append("\nCatagory document counts:\n"); |
| File |
|---|
| ExtractTrainingData.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| Method invocation |
| Chunk |
|---|
| Conflicting content |
|---|
* @return * @throws IOException */ <<<<<<< HEAD public static Map |
| Solution content |
|---|
* @return * @throws IOException */ public static Map |
| File |
|---|
| ExtractTrainingData.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| Method invocation |
| Method signature |
| Variable |
| Chunk |
|---|
| Conflicting content |
|---|
}
return out;
}
<<<<<<< HEAD
/** Close writers opened by {@link #getWriterForCategory(File, String)} */
protected void closeWriters() {
for (SequenceFile.Writer w: writers.values()) {
Closeables.closeQuietly(w);
}
}
/** Append the contents of the specified termVector to a buffer containing a list of terms
*
=======
/**
* Close writers opened by {@link #getWriterForCategory(File, String)}
*/
protected static void closeWriters() {
for (PrintWriter p : trainingWriters.values()) {
p.close();
}
}
/**
* Append the contents of the specified termVector to a buffer containing a list of terms
*
>>>>>>> 678d92be45331bdd660b941831296676aa1233e6
* @param buf
* @param tv
* @param liveDocs |
| Solution content |
|---|
}
return out;
}
/** Close writers opened by {@link #getWriterForCategory(File, String)} */
protected void closeWriters() {
for (SequenceFile.Writer w: writers.values()) {
Closeables.closeQuietly(w);
}
}
/** Append the contents of the specified termVector to a buffer containing a list of terms
*
* @param buf
* @param tv |
| File |
|---|
| ExtractTrainingData.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| Comment |
| Method declaration |
| Chunk |
|---|
| Conflicting content |
|---|
protected static void appendVectorTerms(StringBuilder buf, Terms tv, Bits liveDocs) throws IOException {
if (tv == null) return;
<<<<<<< HEAD
String[] terms = tv.getTerms();
int[] frequencies = tv.getTermFrequencies();
for (int j=0; j < terms.length; j++) {
int freq = frequencies[j];
String term = terms[j];
for (int k=0; k < freq; k++) {
buf.append(term).append(' ');
=======
TermsEnum terms = tv.iterator(null);
BytesRef ref = null;
while ((ref = terms.next()) != null) {
int freq = getFrequency(tv, terms, liveDocs);
for (int k = 0; k < freq; k++) {
buf.append(new String(ref.bytes, ref.offset, ref.length));//TODO: do we need a charset here?
>>>>>>> 678d92be45331bdd660b941831296676aa1233e6
}
}
} |
| Solution content |
|---|
if (tv == null) return;
String[] terms = tv.getTerms();
int[] frequencies = tv.getTermFrequencies();
for (int j=0; j < terms.length; j++) {
int freq = frequencies[j];
String term = terms[j];
for (int k=0; k < freq; k++) {
buf.append(term).append(' ');
}
}
} |
| File |
|---|
| ExtractTrainingData.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| Array access |
| Comment |
| For statement |
| Method invocation |
| Variable |
| While statement |
| Chunk |
|---|
| Conflicting content |
|---|
* @param buf
* @param f
*/
<<<<<<< HEAD
protected static void appendFieldText(StringBuilder buf, Fieldable f) {
=======
protected static void appendFieldText(StringBuilder buf, IndexableField f) {
>>>>>>> 678d92be45331bdd660b941831296676aa1233e6
if (f == null) return;
if (f.binaryValue() != null) return;
if (f.stringValue() == null) return; |
| Solution content |
|---|
* @param buf
* @param f
*/
protected static void appendFieldText(StringBuilder buf, Fieldable f) {
if (f == null) return; |
| File |
|---|
| ExtractTrainingData.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| Method signature |
| Chunk |
|---|
| Conflicting content |
|---|
buf.append(s); } <<<<<<< HEAD /** Split a comma-delimited set of strings into a list * ======= /** * Split a comma-delimited set of strings into a list * >>>>>>> 678d92be45331bdd660b941831296676aa1233e6 * @param input * @return */ |
| Solution content |
|---|
buf.append(s); } /** Split a comma-delimited set of strings into a list * * @param input * @return */ |
| File |
|---|
| ExtractTrainingData.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| Comment |
| Chunk |
|---|
| Conflicting content |
|---|
}
<<<<<<< HEAD
/**
* Defines index management mechanisms for Lucene based classifier model.
=======
/** builda a lucene index suidable for knn based classification. Each category's content is indexed into
* separate documents in the index, and the category that has the haghest count in the tip N hits is
* is the category that is assigned.
* @param inputFiles
* @throws Exception
>>>>>>> 678d92be45331bdd660b941831296676aa1233e6
*/
public static abstract class IndexingCallback implements TwentyNewsgroupsCorpus.Callback {
protected IndexWriter writer; |
| Solution content |
|---|
}
/**
* Defines index management mechanisms for Lucene based classifier model.
*/
public static abstract class IndexingCallback implements TwentyNewsgroupsCorpus.Callback {
protected IndexWriter writer; |
| File |
|---|
| TrainMoreLikeThis.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| Comment |
| Chunk |
|---|
| Conflicting content |
|---|
writer.addDocument(d); // |
| Solution content |
|---|
writer.addDocument(d); // |
| File |
|---|
| TrainMoreLikeThis.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| Comment |
| Method invocation |
| Variable |
| Chunk |
|---|
| Conflicting content |
|---|
log.info("KNN: Added document for category " + category + " named " + inputFile.getName());
}
<<<<<<< HEAD
=======
writer.setCommitData(generateUserData(categories));
writer.commit();
log.info("Knn: Added " + fileCount + " categories in " + (System.currentTimeMillis() - start) + " msec.");
>>>>>>> 678d92be45331bdd660b941831296676aa1233e6
}
/** builds a lucene index suitable for tfidf based classification. Each categories content is indexed into |
| Solution content |
|---|
log.info("KNN: Added document for category " + category + " named " + inputFile.getName());
}
}
/** builds a lucene index suitable for tfidf based classification. Each categories content is indexed into |
| File |
|---|
| TrainMoreLikeThis.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| Method invocation |
| Chunk |
|---|
| Conflicting content |
|---|
*/
public static class TfidfIndexer extends IndexingCallback {
<<<<<<< HEAD
// holds the collected content for each category
final Map |
| Solution content |
|---|
*/
public static class TfidfIndexer extends IndexingCallback {
// holds the collected content for each category
final Map |
| File |
|---|
| TrainMoreLikeThis.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| Attribute |
| Comment |
| For statement |
| Method declaration |
| Method invocation |
| Variable |
| Chunk |
|---|
| Conflicting content |
|---|
d.add(categoryField);
d.add(contentField);
<<<<<<< HEAD
writer.addDocument(d); // |
| Solution content |
|---|
d.add(categoryField);
d.add(contentField);
writer.addDocument(d); // |
| File |
|---|
| TrainMoreLikeThis.java |
| Developer's decision |
|---|
| Version 1 |
| Kind of conflict |
|---|
| Comment |
| Method declaration |
| Method invocation |
| Variable |