| /* |
| * Copyright (C) 2019, Google Inc. and others |
| * |
| * This program and the accompanying materials are made available under the |
| * terms of the Eclipse Distribution License v. 1.0 which is available at |
| * https://www.eclipse.org/org/documents/edl-v10.php. |
| * |
| * SPDX-License-Identifier: BSD-3-Clause |
| */ |
| |
| package org.eclipse.jgit.internal.storage.dfs; |
| |
| import org.eclipse.jgit.internal.storage.dfs.DfsObjDatabase.PackSource; |
| import org.eclipse.jgit.internal.storage.io.BlockSource; |
| import org.eclipse.jgit.internal.storage.pack.PackExt; |
| import org.eclipse.jgit.internal.storage.reftable.ReftableBatchRefUpdate; |
| import org.eclipse.jgit.internal.storage.reftable.ReftableCompactor; |
| import org.eclipse.jgit.internal.storage.reftable.ReftableConfig; |
| import org.eclipse.jgit.internal.storage.reftable.ReftableReader; |
| import org.eclipse.jgit.internal.storage.reftable.ReftableWriter; |
| import org.eclipse.jgit.lib.Ref; |
| import org.eclipse.jgit.transport.ReceiveCommand; |
| |
| import java.io.ByteArrayOutputStream; |
| import java.io.IOException; |
| import java.io.OutputStream; |
| import java.util.ArrayList; |
| import java.util.Collections; |
| import java.util.List; |
| import java.util.Set; |
| |
| import static org.eclipse.jgit.internal.storage.pack.PackExt.REFTABLE; |
| |
| /** |
| * {@link org.eclipse.jgit.lib.BatchRefUpdate} for |
| * {@link org.eclipse.jgit.internal.storage.dfs.DfsReftableDatabase}. |
| */ |
| public class DfsReftableBatchRefUpdate extends ReftableBatchRefUpdate { |
| private static final int AVG_BYTES = 36; |
| |
| private final DfsReftableDatabase refdb; |
| |
| private final DfsObjDatabase odb; |
| |
| /** |
| * Initialize batch update. |
| * |
| * @param refdb |
| * database the update will modify. |
| * @param odb |
| * object database to store the reftable. |
| */ |
| protected DfsReftableBatchRefUpdate(DfsReftableDatabase refdb, |
| DfsObjDatabase odb) { |
| super(refdb, refdb.reftableDatabase, refdb.getLock(), refdb.getRepository()); |
| this.refdb = refdb; |
| this.odb = odb; |
| } |
| |
| @Override |
| protected void applyUpdates(List<Ref> newRefs, List<ReceiveCommand> pending) |
| throws IOException { |
| Set<DfsPackDescription> prune = Collections.emptySet(); |
| DfsPackDescription pack = odb.newPack(PackSource.INSERT); |
| try (DfsOutputStream out = odb.writeFile(pack, REFTABLE)) { |
| ReftableConfig cfg = DfsPackCompactor |
| .configureReftable(refdb.getReftableConfig(), out); |
| |
| ReftableWriter.Stats stats; |
| if (refdb.compactDuringCommit() |
| && newRefs.size() * AVG_BYTES <= cfg.getRefBlockSize() |
| && canCompactTopOfStack(cfg)) { |
| ByteArrayOutputStream tmp = new ByteArrayOutputStream(); |
| ReftableWriter rw = new ReftableWriter(cfg, tmp); |
| write(rw, newRefs, pending); |
| rw.finish(); |
| stats = compactTopOfStack(out, cfg, tmp.toByteArray()); |
| prune = toPruneTopOfStack(); |
| } else { |
| ReftableWriter rw = new ReftableWriter(cfg, out); |
| write(rw, newRefs, pending); |
| rw.finish(); |
| stats = rw.getStats(); |
| } |
| pack.addFileExt(REFTABLE); |
| pack.setReftableStats(stats); |
| } |
| |
| odb.commitPack(Collections.singleton(pack), prune); |
| odb.addReftable(pack, prune); |
| refdb.clearCache(); |
| } |
| |
| private boolean canCompactTopOfStack(ReftableConfig cfg) |
| throws IOException { |
| refdb.getLock().lock(); |
| try { |
| DfsReftableStack stack = refdb.stack(); |
| List<ReftableReader> readers = stack.readers(); |
| if (readers.isEmpty()) { |
| return false; |
| } |
| |
| int lastIdx = readers.size() - 1; |
| DfsReftable last = stack.files().get(lastIdx); |
| DfsPackDescription desc = last.getPackDescription(); |
| if (desc.getPackSource() != PackSource.INSERT |
| || !packOnlyContainsReftable(desc)) { |
| return false; |
| } |
| |
| ReftableReader table = readers.get(lastIdx); |
| int bs = cfg.getRefBlockSize(); |
| return table.size() <= 3 * bs; |
| } finally { |
| refdb.getLock().unlock(); |
| } |
| } |
| |
| private ReftableWriter.Stats compactTopOfStack(OutputStream out, |
| ReftableConfig cfg, byte[] newTable) throws IOException { |
| refdb.getLock().lock(); |
| try { |
| List<ReftableReader> stack = refdb.stack().readers(); |
| |
| ReftableReader last = stack.get(stack.size() - 1); |
| |
| List<ReftableReader> tables = new ArrayList<>(2); |
| tables.add(last); |
| tables.add(new ReftableReader(BlockSource.from(newTable))); |
| |
| ReftableCompactor compactor = new ReftableCompactor(out); |
| compactor.setConfig(cfg); |
| compactor.setIncludeDeletes(true); |
| compactor.addAll(tables); |
| compactor.compact(); |
| return compactor.getStats(); |
| } finally { |
| refdb.getLock().unlock(); |
| } |
| } |
| |
| private Set<DfsPackDescription> toPruneTopOfStack() throws IOException { |
| refdb.getLock().lock(); |
| try { |
| List<DfsReftable> stack = refdb.stack().files(); |
| |
| DfsReftable last = stack.get(stack.size() - 1); |
| return Collections.singleton(last.getPackDescription()); |
| } finally { |
| refdb.getLock().unlock(); |
| } |
| } |
| |
| private boolean packOnlyContainsReftable(DfsPackDescription desc) { |
| for (PackExt ext : PackExt.values()) { |
| if (ext != REFTABLE && desc.hasFileExt(ext)) { |
| return false; |
| } |
| } |
| return true; |
| } |
| } |