|
|
|
@ -44,21 +44,29 @@
|
|
|
|
|
package org.eclipse.jgit.internal.storage.dfs; |
|
|
|
|
|
|
|
|
|
import static org.eclipse.jgit.internal.storage.dfs.DfsObjDatabase.PackSource.COMPACT; |
|
|
|
|
import static org.eclipse.jgit.internal.storage.dfs.DfsObjDatabase.PackSource.GC; |
|
|
|
|
import static org.eclipse.jgit.internal.storage.pack.PackExt.INDEX; |
|
|
|
|
import static org.eclipse.jgit.internal.storage.pack.PackExt.PACK; |
|
|
|
|
import static org.eclipse.jgit.internal.storage.pack.PackExt.REFTABLE; |
|
|
|
|
import static org.eclipse.jgit.internal.storage.pack.StoredObjectRepresentation.PACK_DELTA; |
|
|
|
|
|
|
|
|
|
import java.io.IOException; |
|
|
|
|
import java.util.ArrayList; |
|
|
|
|
import java.util.Collection; |
|
|
|
|
import java.util.Collections; |
|
|
|
|
import java.util.Comparator; |
|
|
|
|
import java.util.HashSet; |
|
|
|
|
import java.util.Iterator; |
|
|
|
|
import java.util.List; |
|
|
|
|
import java.util.Set; |
|
|
|
|
|
|
|
|
|
import org.eclipse.jgit.errors.IncorrectObjectTypeException; |
|
|
|
|
import org.eclipse.jgit.internal.JGitText; |
|
|
|
|
import org.eclipse.jgit.internal.storage.file.PackIndex; |
|
|
|
|
import org.eclipse.jgit.internal.storage.file.PackReverseIndex; |
|
|
|
|
import org.eclipse.jgit.internal.storage.pack.PackWriter; |
|
|
|
|
import org.eclipse.jgit.internal.storage.reftable.ReftableCompactor; |
|
|
|
|
import org.eclipse.jgit.internal.storage.reftable.ReftableConfig; |
|
|
|
|
import org.eclipse.jgit.lib.AnyObjectId; |
|
|
|
|
import org.eclipse.jgit.lib.NullProgressMonitor; |
|
|
|
|
import org.eclipse.jgit.lib.ObjectId; |
|
|
|
@ -89,16 +97,15 @@ import org.eclipse.jgit.util.io.CountingOutputStream;
|
|
|
|
|
*/ |
|
|
|
|
public class DfsPackCompactor { |
|
|
|
|
private final DfsRepository repo; |
|
|
|
|
|
|
|
|
|
private final List<DfsPackFile> srcPacks; |
|
|
|
|
|
|
|
|
|
private final List<DfsReftable> srcReftables; |
|
|
|
|
private final List<ObjectIdSet> exclude; |
|
|
|
|
|
|
|
|
|
private final List<DfsPackDescription> newPacks; |
|
|
|
|
|
|
|
|
|
private final List<PackStatistics> newStats; |
|
|
|
|
private PackStatistics newStats; |
|
|
|
|
private DfsPackDescription outDesc; |
|
|
|
|
|
|
|
|
|
private int autoAddSize; |
|
|
|
|
private ReftableConfig reftableConfig; |
|
|
|
|
|
|
|
|
|
private RevWalk rw; |
|
|
|
|
private RevFlag added; |
|
|
|
@ -114,9 +121,19 @@ public class DfsPackCompactor {
|
|
|
|
|
repo = repository; |
|
|
|
|
autoAddSize = 5 * 1024 * 1024; // 5 MiB
|
|
|
|
|
srcPacks = new ArrayList<>(); |
|
|
|
|
srcReftables = new ArrayList<>(); |
|
|
|
|
exclude = new ArrayList<>(4); |
|
|
|
|
newPacks = new ArrayList<>(1); |
|
|
|
|
newStats = new ArrayList<>(1); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
/** |
|
|
|
|
* @param cfg |
|
|
|
|
* configuration to write a reftable. Reftable compacting is |
|
|
|
|
* disabled (default) when {@code cfg} is {@code null}. |
|
|
|
|
* @return {@code this} |
|
|
|
|
*/ |
|
|
|
|
public DfsPackCompactor setReftableConfig(ReftableConfig cfg) { |
|
|
|
|
reftableConfig = cfg; |
|
|
|
|
return this; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
/** |
|
|
|
@ -137,7 +154,19 @@ public class DfsPackCompactor {
|
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
/** |
|
|
|
|
* Automatically select packs to be included, and add them. |
|
|
|
|
* Add a reftable to be compacted. |
|
|
|
|
* |
|
|
|
|
* @param table |
|
|
|
|
* a reftable to combine. |
|
|
|
|
* @return {@code this} |
|
|
|
|
*/ |
|
|
|
|
public DfsPackCompactor add(DfsReftable table) { |
|
|
|
|
srcReftables.add(table); |
|
|
|
|
return this; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
/** |
|
|
|
|
* Automatically select pack and reftables to be included, and add them. |
|
|
|
|
* <p> |
|
|
|
|
* Packs are selected based on size, smaller packs get included while bigger |
|
|
|
|
* ones are omitted. |
|
|
|
@ -155,6 +184,16 @@ public class DfsPackCompactor {
|
|
|
|
|
else |
|
|
|
|
exclude(pack); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
if (reftableConfig != null) { |
|
|
|
|
for (DfsReftable table : objdb.getReftables()) { |
|
|
|
|
DfsPackDescription d = table.getPackDescription(); |
|
|
|
|
if (d.getPackSource() != GC |
|
|
|
|
&& d.getFileSize(REFTABLE) < autoAddSize) { |
|
|
|
|
add(table); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
return this; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
@ -197,11 +236,30 @@ public class DfsPackCompactor {
|
|
|
|
|
* the packs cannot be compacted. |
|
|
|
|
*/ |
|
|
|
|
public void compact(ProgressMonitor pm) throws IOException { |
|
|
|
|
if (pm == null) |
|
|
|
|
if (pm == null) { |
|
|
|
|
pm = NullProgressMonitor.INSTANCE; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
DfsObjDatabase objdb = repo.getObjectDatabase(); |
|
|
|
|
try (DfsReader ctx = objdb.newReader()) { |
|
|
|
|
if (reftableConfig != null && !srcReftables.isEmpty()) { |
|
|
|
|
compactReftables(ctx); |
|
|
|
|
} |
|
|
|
|
compactPacks(ctx, pm); |
|
|
|
|
|
|
|
|
|
List<DfsPackDescription> commit = getNewPacks(); |
|
|
|
|
Collection<DfsPackDescription> remove = toPrune(); |
|
|
|
|
if (!commit.isEmpty() || !remove.isEmpty()) { |
|
|
|
|
objdb.commitPack(commit, remove); |
|
|
|
|
} |
|
|
|
|
} finally { |
|
|
|
|
rw = null; |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
private void compactPacks(DfsReader ctx, ProgressMonitor pm) |
|
|
|
|
throws IOException, IncorrectObjectTypeException { |
|
|
|
|
DfsObjDatabase objdb = repo.getObjectDatabase(); |
|
|
|
|
PackConfig pc = new PackConfig(repo); |
|
|
|
|
pc.setIndexVersion(2); |
|
|
|
|
pc.setDeltaCompress(false); |
|
|
|
@ -215,40 +273,31 @@ public class DfsPackCompactor {
|
|
|
|
|
|
|
|
|
|
addObjectsToPack(pw, ctx, pm); |
|
|
|
|
if (pw.getObjectCount() == 0) { |
|
|
|
|
List<DfsPackDescription> remove = toPrune(); |
|
|
|
|
if (remove.size() > 0) |
|
|
|
|
objdb.commitPack( |
|
|
|
|
Collections.<DfsPackDescription>emptyList(), |
|
|
|
|
remove); |
|
|
|
|
return; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
boolean rollback = true; |
|
|
|
|
DfsPackDescription pack = objdb.newPack(COMPACT, |
|
|
|
|
estimatePackSize()); |
|
|
|
|
initOutDesc(objdb); |
|
|
|
|
try { |
|
|
|
|
writePack(objdb, pack, pw, pm); |
|
|
|
|
writeIndex(objdb, pack, pw); |
|
|
|
|
writePack(objdb, outDesc, pw, pm); |
|
|
|
|
writeIndex(objdb, outDesc, pw); |
|
|
|
|
|
|
|
|
|
PackStatistics stats = pw.getStatistics(); |
|
|
|
|
pw.close(); |
|
|
|
|
pw = null; |
|
|
|
|
|
|
|
|
|
pack.setPackStats(stats); |
|
|
|
|
objdb.commitPack(Collections.singletonList(pack), toPrune()); |
|
|
|
|
newPacks.add(pack); |
|
|
|
|
newStats.add(stats); |
|
|
|
|
outDesc.setPackStats(stats); |
|
|
|
|
newStats = stats; |
|
|
|
|
rollback = false; |
|
|
|
|
} finally { |
|
|
|
|
if (rollback) |
|
|
|
|
objdb.rollbackPack(Collections.singletonList(pack)); |
|
|
|
|
if (rollback) { |
|
|
|
|
objdb.rollbackPack(Collections.singletonList(outDesc)); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
} finally { |
|
|
|
|
if (pw != null) |
|
|
|
|
if (pw != null) { |
|
|
|
|
pw.close(); |
|
|
|
|
} |
|
|
|
|
} finally { |
|
|
|
|
rw = null; |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
@ -263,27 +312,81 @@ public class DfsPackCompactor {
|
|
|
|
|
return size; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
private void compactReftables(DfsReader ctx) throws IOException { |
|
|
|
|
DfsObjDatabase objdb = repo.getObjectDatabase(); |
|
|
|
|
Collections.sort(srcReftables, objdb.reftableComparator()); |
|
|
|
|
|
|
|
|
|
try (ReftableStack stack = ReftableStack.open(ctx, srcReftables)) { |
|
|
|
|
initOutDesc(objdb); |
|
|
|
|
ReftableCompactor compact = new ReftableCompactor(); |
|
|
|
|
compact.addAll(stack.readers()); |
|
|
|
|
compact.setIncludeDeletes(true); |
|
|
|
|
writeReftable(objdb, outDesc, compact); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
private void initOutDesc(DfsObjDatabase objdb) throws IOException { |
|
|
|
|
if (outDesc == null) { |
|
|
|
|
outDesc = objdb.newPack(COMPACT, estimatePackSize()); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
/** @return all of the source packs that fed into this compaction. */ |
|
|
|
|
public List<DfsPackDescription> getSourcePacks() { |
|
|
|
|
return toPrune(); |
|
|
|
|
public Collection<DfsPackDescription> getSourcePacks() { |
|
|
|
|
Set<DfsPackDescription> src = new HashSet<>(); |
|
|
|
|
for (DfsPackFile pack : srcPacks) { |
|
|
|
|
src.add(pack.getPackDescription()); |
|
|
|
|
} |
|
|
|
|
for (DfsReftable table : srcReftables) { |
|
|
|
|
src.add(table.getPackDescription()); |
|
|
|
|
} |
|
|
|
|
return src; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
/** @return new packs created by this compaction. */ |
|
|
|
|
public List<DfsPackDescription> getNewPacks() { |
|
|
|
|
return newPacks; |
|
|
|
|
return outDesc != null |
|
|
|
|
? Collections.singletonList(outDesc) |
|
|
|
|
: Collections.emptyList(); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
/** @return statistics corresponding to the {@link #getNewPacks()}. */ |
|
|
|
|
public List<PackStatistics> getNewPackStatistics() { |
|
|
|
|
return newStats; |
|
|
|
|
return newStats != null |
|
|
|
|
? Collections.singletonList(newStats) |
|
|
|
|
: Collections.emptyList(); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
private Collection<DfsPackDescription> toPrune() { |
|
|
|
|
Set<DfsPackDescription> packs = new HashSet<>(); |
|
|
|
|
for (DfsPackFile pack : srcPacks) { |
|
|
|
|
packs.add(pack.getPackDescription()); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
Set<DfsPackDescription> reftables = new HashSet<>(); |
|
|
|
|
for (DfsReftable table : srcReftables) { |
|
|
|
|
reftables.add(table.getPackDescription()); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
for (Iterator<DfsPackDescription> i = packs.iterator(); i.hasNext();) { |
|
|
|
|
DfsPackDescription d = i.next(); |
|
|
|
|
if (d.hasFileExt(REFTABLE) && !reftables.contains(d)) { |
|
|
|
|
i.remove(); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
private List<DfsPackDescription> toPrune() { |
|
|
|
|
int cnt = srcPacks.size(); |
|
|
|
|
List<DfsPackDescription> all = new ArrayList<>(cnt); |
|
|
|
|
for (DfsPackFile pack : srcPacks) |
|
|
|
|
all.add(pack.getPackDescription()); |
|
|
|
|
return all; |
|
|
|
|
for (Iterator<DfsPackDescription> i = reftables.iterator(); |
|
|
|
|
i.hasNext();) { |
|
|
|
|
DfsPackDescription d = i.next(); |
|
|
|
|
if (d.hasFileExt(PACK) && !packs.contains(d)) { |
|
|
|
|
i.remove(); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
Set<DfsPackDescription> toPrune = new HashSet<>(); |
|
|
|
|
toPrune.addAll(packs); |
|
|
|
|
toPrune.addAll(reftables); |
|
|
|
|
return toPrune; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
private void addObjectsToPack(PackWriter pw, DfsReader ctx, |
|
|
|
@ -390,6 +493,27 @@ public class DfsPackCompactor {
|
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
private void writeReftable(DfsObjDatabase objdb, DfsPackDescription pack, |
|
|
|
|
ReftableCompactor compact) throws IOException { |
|
|
|
|
try (DfsOutputStream out = objdb.writeFile(pack, REFTABLE)) { |
|
|
|
|
compact.setConfig(configureReftable(reftableConfig, out)); |
|
|
|
|
compact.compact(out); |
|
|
|
|
pack.addFileExt(REFTABLE); |
|
|
|
|
pack.setReftableStats(compact.getStats()); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static ReftableConfig configureReftable(ReftableConfig cfg, |
|
|
|
|
DfsOutputStream out) { |
|
|
|
|
int bs = out.blockSize(); |
|
|
|
|
if (bs > 0) { |
|
|
|
|
cfg = new ReftableConfig(cfg); |
|
|
|
|
cfg.setRefBlockSize(bs); |
|
|
|
|
cfg.setAlignBlocks(true); |
|
|
|
|
} |
|
|
|
|
return cfg; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
private static class ObjectIdWithOffset extends ObjectId { |
|
|
|
|
final long offset; |
|
|
|
|
|
|
|
|
|