| Chunk |
|---|
| Conflicting content |
|---|
import cascading.tuple.Tuple; import cascading.tuple.TupleEntry; import cascading.tuple.TupleEntryCollector; <<<<<<< HEAD import cascading.util.Util; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.mapred.FileOutputFormat; import org.apache.hadoop.mapred.JobConf; import org.apache.hadoop.mapred.OutputCollector; import org.apache.hadoop.mapred.OutputFormat; import org.apache.hadoop.mapred.RecordWriter; import org.apache.hadoop.mapred.Reporter; ======= import org.apache.hadoop.fs.Path; import org.apache.hadoop.mapred.*; >>>>>>> 91f7bc12b3fb66e9156e15d07fd4573d103dda44 import org.apache.log4j.Logger; import java.io.IOException; |
| Solution content |
|---|
import cascading.tuple.Tuple; import cascading.tuple.TupleEntry; import cascading.tuple.TupleEntryCollector; import org.apache.hadoop.fs.Path; import org.apache.hadoop.mapred.*; import org.apache.log4j.Logger; import java.io.IOException; |
| File |
|---|
| TapCollector.java |
| Developer's decision |
|---|
| Version 2 |
| Kind of conflict |
|---|
| Import |
| Chunk |
|---|
| Conflicting content |
|---|
Hadoop18TapUtil.setupTask( conf );
<<<<<<< HEAD
filename = String.format( filenamePattern, conf.getInt( "mapred.task.partition", 0 ) );
conf.set( "mapred.work.output.dir", outputPath.toString() );
if( outputFormat instanceof FileOutputFormat ) // only file based writing uses temp dirs
fileSystem.mkdirs( new Path( conf.get( "mapred.work.output.dir" ), "_temporary" ) );
if( conf.get( "mapred.task.id" ) == null ) // need to stuff a fake id
conf.set( "mapred.task.id", String.format( "attempt_%12.0e_0000_m_000000_0", Math.rint( System.currentTimeMillis() ) ) );
=======
OutputFormat outputFormat = conf.getOutputFormat();
>>>>>>> 91f7bc12b3fb66e9156e15d07fd4573d103dda44
writer = outputFormat.getRecordWriter( null, conf, filename, Reporter.NULL );
} |
| Solution content |
|---|
Hadoop18TapUtil.setupTask( conf );
OutputFormat outputFormat = conf.getOutputFormat();
writer = outputFormat.getRecordWriter( null, conf, filename, Reporter.NULL );
} |
| File |
|---|
| TapCollector.java |
| Developer's decision |
|---|
| Version 2 |
| Kind of conflict |
|---|
| Attribute |
| If statement |
| Method invocation |
| Variable |
| Chunk |
|---|
| Conflicting content |
|---|
}
}
<<<<<<< HEAD
private void moveTaskOutputs() throws IOException
{
Path outputPath = FileOutputFormat.getOutputPath( conf );
String taskIdPath = conf.get( "mapred.task.id" );
Class[] classes = {JobConf.class, String.class};
Object[] parameters = {conf, "_temporary/" + taskIdPath};
Path taskPath = (Path) Util.invokeStaticMethod( FileOutputFormat.class, "getTaskOutputPath", parameters, classes );
taskPath = taskPath.getParent();
FileSystem fileSystem = FileSystem.get( outputPath.toUri(), conf );
if( !fileSystem.getFileStatus( taskPath ).isDir() )
throw new IOException( "path is not a directory: " + taskPath );
FileStatus[] statuses = fileSystem.listStatus( taskPath );
for( FileStatus status : statuses )
{
Path sourcePath = status.getPath();
if( status.isDir() )
throw new IOException( "path is a directory, no support for nested directories: " + sourcePath );
Path targetPath = new Path( outputPath, sourcePath.getName() );
fileSystem.rename( sourcePath, targetPath );
LOG.debug( "moved " + sourcePath + " to " + targetPath );
}
// remove _temporary directory
fileSystem.delete( new Path( conf.get( "mapred.work.output.dir" ), "_temporary" ), true );
}
=======
>>>>>>> 91f7bc12b3fb66e9156e15d07fd4573d103dda44
@Override
public void close()
{ |
| Solution content |
|---|
}
}
@Override
public void close()
{ |
| File |
|---|
| TapCollector.java |
| Developer's decision |
|---|
| Version 2 |
| Kind of conflict |
|---|
| Method declaration |
| Chunk |
|---|
| Conflicting content |
|---|
writer.close( reporter );
<<<<<<< HEAD
if( conf.getOutputFormat() instanceof FileOutputFormat )
moveTaskOutputs();
=======
if( Hadoop18TapUtil.needsTaskCommit( conf ) )
Hadoop18TapUtil.commitTask( conf );
>>>>>>> 91f7bc12b3fb66e9156e15d07fd4573d103dda44
Hadoop18TapUtil.cleanupJob( conf );
} |
| Solution content |
|---|
writer.close( reporter );
if( Hadoop18TapUtil.needsTaskCommit( conf ) )
Hadoop18TapUtil.commitTask( conf );
Hadoop18TapUtil.cleanupJob( conf );
} |
| File |
|---|
| TapCollector.java |
| Developer's decision |
|---|
| Version 2 |
| Kind of conflict |
|---|
| If statement |
| Chunk |
|---|
| Conflicting content |
|---|
package cascading.util; <<<<<<< HEAD import java.io.ByteArrayInputStream; import java.io.ByteArrayOutputStream; import java.io.FileWriter; import java.io.IOException; import java.io.ObjectInputStream; import java.io.ObjectOutputStream; import java.io.PrintStream; import java.io.PrintWriter; import java.io.StringWriter; import java.io.Writer; import java.lang.reflect.Method; import java.util.Collection; import java.util.List; import java.util.Map; ======= >>>>>>> 91f7bc12b3fb66e9156e15d07fd4573d103dda44 import cascading.flow.FlowElement; import cascading.flow.FlowException; import cascading.flow.Scope; |
| Solution content |
|---|
package cascading.util; import cascading.flow.FlowElement; import cascading.flow.FlowException; import cascading.flow.Scope; |
| File |
|---|
| Util.java |
| Developer's decision |
|---|
| Version 2 |
| Kind of conflict |
|---|
| Import |
| Chunk |
|---|
| Conflicting content |
|---|
}
new DOTExporter( vertexIdProvider, vertexNameProvider, edgeNameProvider ).export( writer, graph );
}
<<<<<<< HEAD
=======
public interface RetryOperator |
| Solution content |
|---|
new DOTExporter( vertexIdProvider, vertexNameProvider, edgeNameProvider ).export( writer, graph );
}
public interface RetryOperator |
| File |
|---|
| Util.java |
| Developer's decision |
|---|
| Version 2 |
| Kind of conflict |
|---|
| Interface declaration |
| Method declaration |