| Chunk |
|---|
| Conflicting content |
|---|
*/
public void transferBucketToArchive(Bucket bucket, URI uri) {
try {
<<<<<<< HEAD:src/java/com/splunk/shep/archiver/archive/BucketTransferer.java
archiveFileSystem.putFileAtomically(bucket.getDirectory(), path);
=======
archiveFileSystem.putFile(bucket.getDirectory(), uri);
>>>>>>> d6bd4c739075cd550678709683c17f745301fa25:src/java/com/splunk/shep/archiver/archive/ArchiveBucketTransferer.java
} catch (FileNotFoundException e) {
throw new RuntimeException(e); |
| Solution content |
|---|
*/
public void transferBucketToArchive(Bucket bucket, URI uri) {
try {
archiveFileSystem.putFileAtomically(bucket.getDirectory(), uri);
} catch (FileNotFoundException e) {
throw new RuntimeException(e); |
| File |
|---|
| ArchiveBucketTransferer.java |
| Developer's decision |
|---|
| Manual |
| Kind of conflict |
|---|
| Method invocation |
| Chunk |
|---|
| Conflicting content |
|---|
package com.splunk.shep.archiver.archive;
<<<<<<< HEAD
import java.lang.ref.SoftReference;
import java.net.URI;
import org.apache.hadoop.fs.Path;
// CONFIG This whole class should use MBeans
=======
import java.util.Arrays;
import java.util.List;
>>>>>>> d6bd4c739075cd550678709683c17f745301fa25
public class ArchiveConfiguration {
/** |
| Solution content |
|---|
package com.splunk.shep.archiver.archive;
import java.lang.ref.SoftReference;
import java.net.URI;
import java.util.Arrays;
import java.util.List;
import org.apache.hadoop.fs.Path;
//CONFIG This whole class should use MBeans
public class ArchiveConfiguration {
/** |
| File |
|---|
| ArchiveConfiguration.java |
| Developer's decision |
|---|
| Combination |
| Kind of conflict |
|---|
| Comment |
| Import |
| Chunk |
|---|
| Conflicting content |
|---|
import static org.testng.AssertJUnit.*; <<<<<<< HEAD import org.apache.hadoop.fs.Path; ======= import java.util.Arrays; import java.util.List; >>>>>>> d6bd4c739075cd550678709683c17f745301fa25 import org.testng.annotations.BeforeMethod; import org.testng.annotations.Test; |
| Solution content |
|---|
import static org.testng.AssertJUnit.*; import java.util.Arrays; import java.util.List; import org.apache.hadoop.fs.Path; import org.testng.annotations.BeforeMethod; import org.testng.annotations.Test; |
| File |
|---|
| ArchiveConfigurationTest.java |
| Developer's decision |
|---|
| Concatenation |
| Kind of conflict |
|---|
| Import |
| Chunk |
|---|
| Conflicting content |
|---|
assertNotNull(configuration.getServerName());
}
<<<<<<< HEAD
public void getTmpDirectory_defaultState_isNotNull() {
assertNotNull(configuration.getTmpDirectory());
}
public void getTmpDirectory_defaultState_defaultConfValue() {
assertNotNull(configuration.getTmpDirectory());
}
public void getArchiverHadoopURI_defaultState_isNotNull() {
assertNotNull(configuration.getArchiverHadoopURI());
}
public void getArchiverHadoopURI_defaultState_defaultValue() {
assertEquals(new Path("hdfs://localhost:9000/archiver-tmp"),
configuration.getTmpDirectory());
}
=======
public void getBucketFormatPriority_defaultState_SplunkBucketFormatIsPrioritized() {
List |
| Solution content |
|---|
assertNotNull(configuration.getServerName());
}
public void getTmpDirectory_defaultState_isNotNull() {
assertNotNull(configuration.getTmpDirectory());
}
public void getTmpDirectory_defaultState_defaultConfValue() {
assertNotNull(configuration.getTmpDirectory());
}
public void getArchiverHadoopURI_defaultState_isNotNull() {
assertNotNull(configuration.getArchiverHadoopURI());
}
public void getArchiverHadoopURI_defaultState_defaultValue() {
assertEquals(new Path("hdfs://localhost:9000/archiver-tmp"),
configuration.getTmpDirectory());
}
public void getBucketFormatPriority_defaultState_SplunkBucketFormatIsPrioritized() {
List |
| File |
|---|
| ArchiveConfigurationTest.java |
| Developer's decision |
|---|
| Concatenation |
| Kind of conflict |
|---|
| Method declaration |