9
0
mirror of https://github.com/BX-Team/DivineMC.git synced 2025-12-19 14:59:25 +00:00

add back old linear impl; switch to AT's

This commit is contained in:
NONPLAYT
2025-07-13 01:04:45 +03:00
parent 2c5f1ee9f3
commit 650d06ae5c
7 changed files with 648 additions and 48 deletions

View File

@@ -15,6 +15,11 @@ public net.minecraft.world.entity.ai.sensing.Sensor timeToTick
public net.minecraft.world.entity.animal.armadillo.Armadillo scuteTime
public net.minecraft.world.entity.animal.frog.Tadpole getTicksLeftUntilAdult()I
public net.minecraft.world.level.chunk.PaletteResize
public net.minecraft.world.level.chunk.storage.RegionFile getOversizedData(II)Lnet/minecraft/nbt/CompoundTag;
public net.minecraft.world.level.chunk.storage.RegionFile isOversized(II)Z
public net.minecraft.world.level.chunk.storage.RegionFile recalculateHeader()Z
public net.minecraft.world.level.chunk.storage.RegionFile setOversized(IIZ)V
public net.minecraft.world.level.chunk.storage.RegionFile write(Lnet/minecraft/world/level/ChunkPos;Ljava/nio/ByteBuffer;)V
public net.minecraft.world.level.entity.EntityTickList entities
public net.minecraft.world.level.levelgen.DensityFunctions$BlendAlpha
public net.minecraft.world.level.levelgen.DensityFunctions$BlendDensity

View File

@@ -72,7 +72,7 @@
}
}
val log4jPlugins = sourceSets.create("log4jPlugins") {
@@ -156,10 +_,20 @@
@@ -156,10 +_,21 @@
}
dependencies {
@@ -86,6 +86,7 @@
+ }
+ implementation("com.github.luben:zstd-jni:1.5.7-3")
+ implementation("org.lz4:lz4-java:1.8.0")
+ implementation("net.openhft:zero-allocation-hashing:0.16")
+ // DivineMC end - Dependencies
+
implementation("ca.spottedleaf:concurrentutil:0.0.3")

View File

@@ -1,7 +1,7 @@
From 0000000000000000000000000000000000000000 Mon Sep 17 00:00:00 2001
From: NONPLAYT <76615486+NONPLAYT@users.noreply.github.com>
Date: Fri, 11 Jul 2025 21:47:45 +0300
Subject: [PATCH] Buffered Linear region format
Subject: [PATCH] Linear region file format
diff --git a/ca/spottedleaf/moonrise/patches/chunk_system/io/ChunkSystemRegionFileStorage.java b/ca/spottedleaf/moonrise/patches/chunk_system/io/ChunkSystemRegionFileStorage.java
@@ -126,7 +126,7 @@ index 79d57ca8a7870a02e95562d89cbd4341d8282660..1156772217b139d54266f470b18d4a98
class PoiUpgrader extends WorldUpgrader.SimpleRegionStorageUpgrader {
diff --git a/net/minecraft/world/level/chunk/storage/RegionFile.java b/net/minecraft/world/level/chunk/storage/RegionFile.java
index 22f3aa1674664906e8ec45372d758d79017e3987..55eaf7a5d4ceb957717298991fecce0b81c0f377 100644
index ae0a893498d0bfe90c14508f15b431d4885e06ff..00656cf8634e06f7ce1067ef7ba44edfb4519be3 100644
--- a/net/minecraft/world/level/chunk/storage/RegionFile.java
+++ b/net/minecraft/world/level/chunk/storage/RegionFile.java
@@ -22,7 +22,7 @@ import net.minecraft.util.profiling.jfr.JvmProfiler;
@@ -138,24 +138,6 @@ index 22f3aa1674664906e8ec45372d758d79017e3987..55eaf7a5d4ceb957717298991fecce0b
private static final Logger LOGGER = LogUtils.getLogger();
public static final int MAX_CHUNK_SIZE = 500 * 1024 * 1024; // Paper - don't write garbage data to disk if writing serialization fails
private static final int SECTOR_BYTES = 4096;
@@ -130,7 +130,7 @@ public class RegionFile implements AutoCloseable, ca.spottedleaf.moonrise.patche
return this.recalculateCount.get();
}
- boolean recalculateHeader() throws IOException {
+ public boolean recalculateHeader() throws IOException { // DivineMC - Buffered Linear region format
if (!this.canRecalcHeader) {
return false;
}
@@ -794,7 +794,7 @@ public class RegionFile implements AutoCloseable, ca.spottedleaf.moonrise.patche
}
}
- protected synchronized void write(ChunkPos chunkPos, ByteBuffer chunkData) throws IOException {
+ public synchronized void write(ChunkPos chunkPos, ByteBuffer chunkData) throws IOException { // DivineMC - Buffered Linear region format
int offsetIndex = getOffsetIndex(chunkPos);
int i = this.offsets.get(offsetIndex);
int sectorNumber = getSectorNumber(i);
@@ -912,7 +912,7 @@ public class RegionFile implements AutoCloseable, ca.spottedleaf.moonrise.patche
}
@@ -165,29 +147,6 @@ index 22f3aa1674664906e8ec45372d758d79017e3987..55eaf7a5d4ceb957717298991fecce0b
regionFile.write(this.pos, ByteBuffer.wrap(this.buf, 0, this.count));
}
// Paper end - rewrite chunk system
@@ -978,11 +978,11 @@ public class RegionFile implements AutoCloseable, ca.spottedleaf.moonrise.patche
return (x & 31) + (z & 31) * 32;
}
- synchronized boolean isOversized(int x, int z) {
+ public synchronized boolean isOversized(int x, int z) { // DivineMC - Buffered Linear region format
return this.oversized[getChunkIndex(x, z)] == 1;
}
- synchronized void setOversized(int x, int z, boolean oversized) throws IOException {
+ public synchronized void setOversized(int x, int z, boolean oversized) throws IOException { // DivineMC - Buffered Linear region format
final int offset = getChunkIndex(x, z);
boolean previous = this.oversized[offset] == 1;
this.oversized[offset] = (byte) (oversized ? 1 : 0);
@@ -1021,7 +1021,7 @@ public class RegionFile implements AutoCloseable, ca.spottedleaf.moonrise.patche
return this.path.getParent().resolve(this.path.getFileName().toString().replaceAll("\\.mca$", "") + "_oversized_" + x + "_" + z + ".nbt");
}
- synchronized net.minecraft.nbt.CompoundTag getOversizedData(int x, int z) throws IOException {
+ public synchronized net.minecraft.nbt.CompoundTag getOversizedData(int x, int z) throws IOException {
Path file = getOversizedFile(x, z);
try (DataInputStream out = new DataInputStream(new java.io.BufferedInputStream(new java.util.zip.InflaterInputStream(Files.newInputStream(file))))) {
return net.minecraft.nbt.NbtIo.read((java.io.DataInput) out);
diff --git a/net/minecraft/world/level/chunk/storage/RegionFileStorage.java b/net/minecraft/world/level/chunk/storage/RegionFileStorage.java
index 8d1174f25e0e90d0533970f4ddd8448442024936..ee797d6b3cd898cba1abd3422cb54b17eb4a639f 100644
--- a/net/minecraft/world/level/chunk/storage/RegionFileStorage.java

View File

@@ -12,6 +12,7 @@ import org.bukkit.configuration.MemoryConfiguration;
import org.bxteam.divinemc.config.annotations.Experimental;
import org.bxteam.divinemc.entity.pathfinding.PathfindTaskRejectPolicy;
import org.bxteam.divinemc.region.EnumRegionFileExtension;
import org.bxteam.divinemc.region.type.LinearRegionFile;
import org.bxteam.divinemc.server.network.AsyncJoinHandler;
import org.jetbrains.annotations.Nullable;
import org.simpleyaml.configuration.comments.CommentType;
@@ -599,6 +600,9 @@ public class DivineConfig {
// Region Format
public static EnumRegionFileExtension regionFileType = EnumRegionFileExtension.MCA;
public static int linearCompressionLevel = 1;
public static int linearIoThreadCount = 6;
public static int linearIoFlushDelayMs = 100;
public static boolean linearUseVirtualThreads = true;
// Sentry
public static String sentryDsn = "";
@@ -649,18 +653,39 @@ public class DivineConfig {
}
private static void regionFileExtension() {
regionFileType = EnumRegionFileExtension.fromString(getString(ConfigCategory.MISC.key("region-format.type"), regionFileType.toString(),
EnumRegionFileExtension configuredType = EnumRegionFileExtension.fromString(getString(ConfigCategory.MISC.key("region-format.type"), regionFileType.toString(),
"The type of region file format to use for storing chunk data.",
"Valid values:",
" - MCA: Default Minecraft region file format",
" - B_LINEAR: Buffered region file format"));
" - LINEAR: Linear region file format V2",
" - B_LINEAR: Buffered region file format (just uses Zstd)"));
if (configuredType != null) {
regionFileType = configuredType;
} else {
LOGGER.warn("Invalid region file type: {}, resetting to default (MCA)", getString(ConfigCategory.MISC.key("region-format.type"), regionFileType.toString()));
regionFileType = EnumRegionFileExtension.MCA;
}
linearCompressionLevel = getInt(ConfigCategory.MISC.key("region-format.compression-level"), linearCompressionLevel,
"The compression level to use for the linear region file format.");
linearIoThreadCount = getInt(ConfigCategory.MISC.key("region-format.linear-io-thread-count"), linearIoThreadCount,
"The number of threads to use for IO operations.");
linearIoFlushDelayMs = getInt(ConfigCategory.MISC.key("region-format.linear-io-flush-delay-ms"), linearIoFlushDelayMs,
"The delay in milliseconds to wait before flushing IO operations.");
linearUseVirtualThreads = getBoolean(ConfigCategory.MISC.key("region-format.linear-use-virtual-threads"), linearUseVirtualThreads,
"Whether to use virtual threads for IO operations that was introduced in Java 21.");
if (linearCompressionLevel > 22 || linearCompressionLevel < 1) {
LOGGER.warn("Invalid linear compression level: {}, resetting to default (1)", linearCompressionLevel);
linearCompressionLevel = 1;
}
if (regionFileType == EnumRegionFileExtension.LINEAR) {
LinearRegionFile.SAVE_DELAY_MS = linearIoFlushDelayMs;
LinearRegionFile.SAVE_THREAD_MAX_COUNT = linearIoThreadCount;
LinearRegionFile.USE_VIRTUAL_THREAD = linearUseVirtualThreads;
}
}
private static void sentrySettings() {

View File

@@ -2,10 +2,13 @@ package org.bxteam.divinemc.region;
import net.minecraft.world.level.chunk.storage.RegionFile;
import org.bxteam.divinemc.config.DivineConfig;
import org.bxteam.divinemc.region.type.BufferedRegionFile;
import org.bxteam.divinemc.region.type.LinearRegionFile;
import org.jetbrains.annotations.Nullable;
public enum EnumRegionFileExtension {
MCA("mca", "mca", (info) -> new RegionFile(info.info(), info.filePath(), info.folder(), info.sync())),
LINEAR("linear", "linear", (info) -> new LinearRegionFile(info.info(), info.filePath(), info.folder(), info.sync(), DivineConfig.MiscCategory.linearCompressionLevel)),
B_LINEAR("b_linear", "b_linear", (info) -> new BufferedRegionFile(info.filePath(), DivineConfig.MiscCategory.linearCompressionLevel));
private final String name;

View File

@@ -1,10 +1,12 @@
package org.bxteam.divinemc.region;
package org.bxteam.divinemc.region.type;
import ca.spottedleaf.moonrise.patches.chunk_system.io.MoonriseRegionFileIO;
import net.jpountz.xxhash.XXHash32;
import net.jpountz.xxhash.XXHashFactory;
import net.minecraft.nbt.CompoundTag;
import net.minecraft.world.level.ChunkPos;
import org.bxteam.divinemc.region.BufferReleaser;
import org.bxteam.divinemc.region.IRegionFile;
import org.jetbrains.annotations.NotNull;
import org.jetbrains.annotations.Nullable;

View File

@@ -0,0 +1,605 @@
package org.bxteam.divinemc.region.type;
import ca.spottedleaf.moonrise.patches.chunk_system.io.MoonriseRegionFileIO;
import com.github.luben.zstd.ZstdInputStream;
import com.github.luben.zstd.ZstdOutputStream;
import net.jpountz.lz4.LZ4Compressor;
import net.jpountz.lz4.LZ4Factory;
import net.jpountz.lz4.LZ4FastDecompressor;
import net.minecraft.nbt.CompoundTag;
import net.minecraft.server.MinecraftServer;
import net.minecraft.world.level.ChunkPos;
import net.minecraft.world.level.chunk.storage.RegionFileVersion;
import net.minecraft.world.level.chunk.storage.RegionStorageInfo;
import net.openhft.hashing.LongHashFunction;
import org.apache.logging.log4j.LogManager;
import org.apache.logging.log4j.Logger;
import org.bxteam.divinemc.region.IRegionFile;
import org.checkerframework.checker.nullness.qual.Nullable;
import java.io.BufferedOutputStream;
import java.io.ByteArrayInputStream;
import java.io.ByteArrayOutputStream;
import java.io.DataInputStream;
import java.io.DataOutputStream;
import java.io.File;
import java.io.FileOutputStream;
import java.io.IOException;
import java.io.InputStream;
import java.nio.ByteBuffer;
import java.nio.file.Files;
import java.nio.file.Path;
import java.nio.file.StandardCopyOption;
import java.util.concurrent.TimeUnit;
import java.util.concurrent.locks.LockSupport;
import java.util.concurrent.locks.ReentrantLock;
public class LinearRegionFile implements IRegionFile {
private static final long SUPERBLOCK = 0xc3ff13183cca9d9aL;
private static final byte VERSION = 3;
private static final int HEADER_SIZE = 27;
private static final int FOOTER_SIZE = 8;
private static final Logger LOGGER = LogManager.getLogger(LinearRegionFile.class.getSimpleName());
private static final Object saveLock = new Object();
public static final int MAX_CHUNK_SIZE = 500 * 1024 * 1024;
public static int SAVE_THREAD_MAX_COUNT = 6;
public static int SAVE_DELAY_MS = 100;
public static boolean USE_VIRTUAL_THREAD = true;
private static int activeSaveThreads = 0;
public final ReentrantLock fileLock = new ReentrantLock(true);
public Path regionFile;
public boolean regionFileOpen = false;
private final byte[][] buffer = new byte[1024][];
private final int[] bufferUncompressedSize = new int[1024];
private final long[] chunkTimestamps = new long[1024];
private final Object markedToSaveLock = new Object();
private final LZ4Compressor compressor;
private final LZ4FastDecompressor decompressor;
private final int compressionLevel;
private final Thread bindThread;
private final java.util.concurrent.atomic.AtomicInteger recalculateCount = new java.util.concurrent.atomic.AtomicInteger();
private byte[][] bucketBuffers;
private boolean markedToSave = false;
private boolean close = false;
private int gridSize = 8;
private int bucketSize = 4;
public LinearRegionFile(RegionStorageInfo storageKey, Path directory, Path path, boolean dsync, int compressionLevel) throws IOException {
this(storageKey, directory, path, RegionFileVersion.getCompressionFormat(), dsync, compressionLevel);
}
public LinearRegionFile(RegionStorageInfo storageKey, Path path, Path directory, RegionFileVersion compressionFormat, boolean dsync, int compressionLevel) throws IOException {
Runnable flushCheck = () -> {
while (!close) {
synchronized (saveLock) {
if (markedToSave && activeSaveThreads < SAVE_THREAD_MAX_COUNT) {
activeSaveThreads++;
Runnable flushOperation = () -> {
try {
flush();
} catch (IOException ex) {
LOGGER.error("Region file {} flush failed", this.regionFile.toAbsolutePath(), ex);
} finally {
synchronized (saveLock) {
activeSaveThreads--;
}
}
};
Thread saveThread = USE_VIRTUAL_THREAD ?
Thread.ofVirtual().name("Linear IO - " + LinearRegionFile.this.hashCode()).unstarted(flushOperation) :
Thread.ofPlatform().name("Linear IO - " + LinearRegionFile.this.hashCode()).unstarted(flushOperation);
saveThread.setPriority(Thread.NORM_PRIORITY - 3);
saveThread.start();
}
}
LockSupport.parkNanos(TimeUnit.MILLISECONDS.toNanos(SAVE_DELAY_MS));
}
};
this.bindThread = USE_VIRTUAL_THREAD ? Thread.ofVirtual().unstarted(flushCheck) : Thread.ofPlatform().unstarted(flushCheck);
this.bindThread.setName("Linear IO Schedule - " + this.hashCode());
this.regionFile = path;
this.compressionLevel = compressionLevel;
this.compressor = LZ4Factory.fastestInstance().fastCompressor();
this.decompressor = LZ4Factory.fastestInstance().fastDecompressor();
}
public Path getRegionFile() {
return this.regionFile;
}
public ReentrantLock getFileLock() {
return this.fileLock;
}
public Path getPath() {
return this.regionFile;
}
public int getRecalculateCount() {
return this.recalculateCount.get();
}
public boolean recalculateHeader() {
return false;
}
private int chunkToBucketIdx(int chunkX, int chunkZ) {
int bx = chunkX / bucketSize, bz = chunkZ / bucketSize;
return bx * gridSize + bz;
}
private void openBucket(int chunkX, int chunkZ) {
chunkX = Math.floorMod(chunkX, 32);
chunkZ = Math.floorMod(chunkZ, 32);
int idx = chunkToBucketIdx(chunkX, chunkZ);
if (bucketBuffers == null) return;
if (bucketBuffers[idx] != null) {
try {
ByteArrayInputStream bucketByteStream = new ByteArrayInputStream(bucketBuffers[idx]);
ZstdInputStream zstdStream = new ZstdInputStream(bucketByteStream);
ByteBuffer bucketBuffer = ByteBuffer.wrap(zstdStream.readAllBytes());
int bx = chunkX / bucketSize, bz = chunkZ / bucketSize;
for (int cx = 0; cx < 32 / gridSize; cx++) {
for (int cz = 0; cz < 32 / gridSize; cz++) {
int chunkIndex = (bx * (32 / gridSize) + cx) + (bz * (32 / gridSize) + cz) * 32;
int chunkSize = bucketBuffer.getInt();
long timestamp = bucketBuffer.getLong();
this.chunkTimestamps[chunkIndex] = timestamp;
if (chunkSize > 0) {
byte[] chunkData = new byte[chunkSize - 8];
bucketBuffer.get(chunkData);
int maxCompressedLength = this.compressor.maxCompressedLength(chunkData.length);
byte[] compressed = new byte[maxCompressedLength];
int compressedLength = this.compressor.compress(chunkData, 0, chunkData.length, compressed, 0, maxCompressedLength);
byte[] finalCompressed = new byte[compressedLength];
System.arraycopy(compressed, 0, finalCompressed, 0, compressedLength);
if (chunkX == cx && chunkZ == cz) {
this.buffer[chunkIndex] = finalCompressed;
this.bufferUncompressedSize[chunkIndex] = chunkData.length;
return;
}
this.buffer[chunkIndex] = finalCompressed;
this.bufferUncompressedSize[chunkIndex] = chunkData.length;
}
}
}
} catch (IOException ex) {
LOGGER.error("Region file corrupted: {} bucket: {}", regionFile, idx);
MinecraftServer.getServer().safeShutdown(true, false);
}
bucketBuffers[idx] = null;
}
}
private synchronized void openRegionFile() {
if (regionFileOpen) return;
regionFileOpen = true;
File regionFile = new File(this.regionFile.toString());
if(!regionFile.canRead()) {
this.bindThread.start();
return;
}
try {
byte[] fileContent = Files.readAllBytes(this.regionFile);
ByteBuffer buffer = ByteBuffer.wrap(fileContent);
long superBlock = buffer.getLong();
if (superBlock != SUPERBLOCK)
throw new RuntimeException("Invalid superblock: " + superBlock + " file " + this.regionFile);
byte version = buffer.get();
if (version == 1 || version == 2) {
parseLinearV1(buffer);
} else if (version == 3) {
parseLinearV2(buffer);
} else {
throw new RuntimeException("Invalid version: " + version + " file " + this.regionFile);
}
this.bindThread.start();
} catch (IOException e) {
throw new RuntimeException("Failed to open region file " + this.regionFile, e);
}
}
private void parseLinearV1(ByteBuffer buffer) throws IOException {
final int HEADER_SIZE = 32;
final int FOOTER_SIZE = 8;
// Skip newestTimestamp (Long) + Compression level (Byte) + Chunk count (Short): Unused.
buffer.position(buffer.position() + 11);
int dataCount = buffer.getInt();
long fileLength = this.regionFile.toFile().length();
if (fileLength != HEADER_SIZE + dataCount + FOOTER_SIZE) {
throw new IOException("Invalid file length: " + this.regionFile + " " + fileLength + " " + (HEADER_SIZE + dataCount + FOOTER_SIZE));
}
buffer.position(buffer.position() + 8); // Skip data hash (Long): Unused.
byte[] rawCompressed = new byte[dataCount];
buffer.get(rawCompressed);
ByteArrayInputStream byteArrayInputStream = new ByteArrayInputStream(rawCompressed);
ZstdInputStream zstdInputStream = new ZstdInputStream(byteArrayInputStream);
ByteBuffer decompressedBuffer = ByteBuffer.wrap(zstdInputStream.readAllBytes());
int[] starts = new int[1024];
for (int i = 0; i < 1024; i++) {
starts[i] = decompressedBuffer.getInt();
decompressedBuffer.getInt(); // Skip timestamps (Int): Unused.
}
for (int i = 0; i < 1024; i++) {
if (starts[i] > 0) {
int size = starts[i];
byte[] chunkData = new byte[size];
decompressedBuffer.get(chunkData);
int maxCompressedLength = this.compressor.maxCompressedLength(size);
byte[] compressed = new byte[maxCompressedLength];
int compressedLength = this.compressor.compress(chunkData, 0, size, compressed, 0, maxCompressedLength);
byte[] finalCompressed = new byte[compressedLength];
System.arraycopy(compressed, 0, finalCompressed, 0, compressedLength);
this.buffer[i] = finalCompressed;
this.bufferUncompressedSize[i] = size;
this.chunkTimestamps[i] = getTimestamp(); // Use current timestamp as we don't have the original
}
}
}
private void parseLinearV2(ByteBuffer buffer) throws IOException {
buffer.getLong(); // Skip newestTimestamp (Long)
gridSize = buffer.get();
if (gridSize != 1 && gridSize != 2 && gridSize != 4 && gridSize != 8 && gridSize != 16 && gridSize != 32)
throw new RuntimeException("Invalid grid size: " + gridSize + " file " + this.regionFile);
bucketSize = 32 / gridSize;
buffer.getInt(); // Skip region_x (Int)
buffer.getInt(); // Skip region_z (Int)
boolean[] chunkExistenceBitmap = deserializeExistenceBitmap(buffer);
while (true) {
byte featureNameLength = buffer.get();
if (featureNameLength == 0) break;
byte[] featureNameBytes = new byte[featureNameLength];
buffer.get(featureNameBytes);
String featureName = new String(featureNameBytes);
int featureValue = buffer.getInt();
// System.out.println("NBT Feature: " + featureName + " = " + featureValue);
}
int[] bucketSizes = new int[gridSize * gridSize];
byte[] bucketCompressionLevels = new byte[gridSize * gridSize];
long[] bucketHashes = new long[gridSize * gridSize];
for (int i = 0; i < gridSize * gridSize; i++) {
bucketSizes[i] = buffer.getInt();
bucketCompressionLevels[i] = buffer.get();
bucketHashes[i] = buffer.getLong();
}
bucketBuffers = new byte[gridSize * gridSize][];
for (int i = 0; i < gridSize * gridSize; i++) {
if (bucketSizes[i] > 0) {
bucketBuffers[i] = new byte[bucketSizes[i]];
buffer.get(bucketBuffers[i]);
long rawHash = LongHashFunction.xx().hashBytes(bucketBuffers[i]);
if (rawHash != bucketHashes[i]) throw new IOException("Region file hash incorrect " + this.regionFile);
}
}
long footerSuperBlock = buffer.getLong();
if (footerSuperBlock != SUPERBLOCK)
throw new IOException("Footer superblock invalid " + this.regionFile);
}
@Override
public MoonriseRegionFileIO.RegionDataController.WriteData moonrise$startWrite(CompoundTag data, ChunkPos pos) throws IOException {
final DataOutputStream out = this.getChunkDataOutputStream(pos);
return new ca.spottedleaf.moonrise.patches.chunk_system.io.MoonriseRegionFileIO.RegionDataController.WriteData(
data, ca.spottedleaf.moonrise.patches.chunk_system.io.MoonriseRegionFileIO.RegionDataController.WriteData.WriteResult.WRITE,
out, regionFile -> out.close()
);
}
private synchronized void markToSave() {
synchronized(markedToSaveLock) {
markedToSave = true;
}
}
private synchronized boolean isMarkedToSave() {
synchronized(markedToSaveLock) {
if(markedToSave) {
markedToSave = false;
return true;
}
return false;
}
}
public synchronized boolean doesChunkExist(ChunkPos pos) throws Exception {
openRegionFile();
throw new Exception("doesChunkExist is a stub");
}
public synchronized boolean hasChunk(ChunkPos pos) {
openRegionFile();
openBucket(pos.x, pos.z);
return this.bufferUncompressedSize[getChunkIndex(pos.x, pos.z)] > 0;
}
public synchronized void write(ChunkPos pos, ByteBuffer buffer) {
openRegionFile();
openBucket(pos.x, pos.z);
try {
byte[] b = toByteArray(new ByteArrayInputStream(buffer.array()));
int uncompressedSize = b.length;
if (uncompressedSize > MAX_CHUNK_SIZE) {
LOGGER.error("Chunk dupe attempt {}", this.regionFile);
clear(pos);
} else {
int maxCompressedLength = this.compressor.maxCompressedLength(b.length);
byte[] compressed = new byte[maxCompressedLength];
int compressedLength = this.compressor.compress(b, 0, b.length, compressed, 0, maxCompressedLength);
b = new byte[compressedLength];
System.arraycopy(compressed, 0, b, 0, compressedLength);
int index = getChunkIndex(pos.x, pos.z);
this.buffer[index] = b;
this.chunkTimestamps[index] = getTimestamp();
this.bufferUncompressedSize[getChunkIndex(pos.x, pos.z)] = uncompressedSize;
}
} catch (IOException e) {
LOGGER.error("Chunk write IOException {} {}", e, this.regionFile);
}
markToSave();
}
public DataOutputStream getChunkDataOutputStream(ChunkPos pos) {
openRegionFile();
openBucket(pos.x, pos.z);
return new DataOutputStream(new BufferedOutputStream(new LinearRegionFile.ChunkBuffer(pos)));
}
@Nullable
public synchronized DataInputStream getChunkDataInputStream(ChunkPos pos) {
openRegionFile();
openBucket(pos.x, pos.z);
if(this.bufferUncompressedSize[getChunkIndex(pos.x, pos.z)] != 0) {
byte[] content = new byte[bufferUncompressedSize[getChunkIndex(pos.x, pos.z)]];
this.decompressor.decompress(this.buffer[getChunkIndex(pos.x, pos.z)], 0, content, 0, bufferUncompressedSize[getChunkIndex(pos.x, pos.z)]);
return new DataInputStream(new ByteArrayInputStream(content));
}
return null;
}
public synchronized void clear(ChunkPos pos) {
openRegionFile();
openBucket(pos.x, pos.z);
int i = getChunkIndex(pos.x, pos.z);
this.buffer[i] = null;
this.bufferUncompressedSize[i] = 0;
this.chunkTimestamps[i] = 0;
markToSave();
}
public synchronized void close() throws IOException {
openRegionFile();
close = true;
try {
flush();
} catch(IOException e) {
throw new IOException("Region flush IOException " + e + " " + this.regionFile);
}
}
public synchronized void flush() throws IOException {
if (!isMarkedToSave()) return;
openRegionFile();
long timestamp = getTimestamp();
long writeStart = System.nanoTime();
File tempFile = new File(regionFile.toString() + ".tmp");
FileOutputStream fileStream = new FileOutputStream(tempFile);
DataOutputStream dataStream = new DataOutputStream(fileStream);
dataStream.writeLong(SUPERBLOCK);
dataStream.writeByte(VERSION);
dataStream.writeLong(timestamp);
dataStream.writeByte(gridSize);
String fileName = regionFile.getFileName().toString();
String[] parts = fileName.split("\\.");
int regionX = 0;
int regionZ = 0;
try {
if (parts.length >= 4) {
regionX = Integer.parseInt(parts[1]);
regionZ = Integer.parseInt(parts[2]);
} else {
LOGGER.warn("Unexpected file name format: {}", fileName);
}
} catch (NumberFormatException e) {
LOGGER.error("Failed to parse region coordinates from file name: {}", fileName, e);
}
dataStream.writeInt(regionX);
dataStream.writeInt(regionZ);
boolean[] chunkExistenceBitmap = new boolean[1024];
for (int i = 0; i < 1024; i++) {
chunkExistenceBitmap[i] = (this.bufferUncompressedSize[i] > 0);
}
writeSerializedExistenceBitmap(dataStream, chunkExistenceBitmap);
writeNBTFeatures(dataStream);
int bucketMisses = 0;
byte[][] buckets = new byte[gridSize * gridSize][];
for (int bx = 0; bx < gridSize; bx++) {
for (int bz = 0; bz < gridSize; bz++) {
if (bucketBuffers != null && bucketBuffers[bx * gridSize + bz] != null) {
buckets[bx * gridSize + bz] = bucketBuffers[bx * gridSize + bz];
continue;
}
bucketMisses++;
ByteArrayOutputStream bucketStream = new ByteArrayOutputStream();
ZstdOutputStream zstdStream = new ZstdOutputStream(bucketStream, this.compressionLevel);
DataOutputStream bucketDataStream = new DataOutputStream(zstdStream);
boolean hasData = false;
for (int cx = 0; cx < 32 / gridSize; cx++) {
for (int cz = 0; cz < 32 / gridSize; cz++) {
int chunkIndex = (bx * 32 / gridSize + cx) + (bz * 32 / gridSize + cz) * 32;
if (this.bufferUncompressedSize[chunkIndex] > 0) {
hasData = true;
byte[] chunkData = new byte[this.bufferUncompressedSize[chunkIndex]];
this.decompressor.decompress(this.buffer[chunkIndex], 0, chunkData, 0, this.bufferUncompressedSize[chunkIndex]);
bucketDataStream.writeInt(chunkData.length + 8);
bucketDataStream.writeLong(this.chunkTimestamps[chunkIndex]);
bucketDataStream.write(chunkData);
} else {
bucketDataStream.writeInt(0);
bucketDataStream.writeLong(this.chunkTimestamps[chunkIndex]);
}
}
}
bucketDataStream.close();
if (hasData) {
buckets[bx * gridSize + bz] = bucketStream.toByteArray();
}
}
}
for (int i = 0; i < gridSize * gridSize; i++) {
dataStream.writeInt(buckets[i] != null ? buckets[i].length : 0);
dataStream.writeByte(this.compressionLevel);
long rawHash = 0;
if (buckets[i] != null) {
rawHash = LongHashFunction.xx().hashBytes(buckets[i]);
}
dataStream.writeLong(rawHash);
}
for (int i = 0; i < gridSize * gridSize; i++) {
if (buckets[i] != null) {
dataStream.write(buckets[i]);
}
}
dataStream.writeLong(SUPERBLOCK);
dataStream.flush();
fileStream.getFD().sync();
fileStream.getChannel().force(true); // Ensure atomicity on Btrfs
dataStream.close();
fileStream.close();
Files.move(tempFile.toPath(), this.regionFile, StandardCopyOption.REPLACE_EXISTING);
}
private void writeNBTFeatures(DataOutputStream dataStream) throws IOException {
// writeNBTFeature(dataStream, "example", 1);
dataStream.writeByte(0); // End of NBT features
}
private void writeNBTFeature(DataOutputStream dataStream, String featureName, int featureValue) throws IOException {
byte[] featureNameBytes = featureName.getBytes();
dataStream.writeByte(featureNameBytes.length);
dataStream.write(featureNameBytes);
dataStream.writeInt(featureValue);
}
private boolean[] deserializeExistenceBitmap(ByteBuffer buffer) {
boolean[] result = new boolean[1024];
for (int i = 0; i < 128; i++) {
byte b = buffer.get();
for (int j = 0; j < 8; j++) {
result[i * 8 + j] = ((b >> (7 - j)) & 1) == 1;
}
}
return result;
}
private void writeSerializedExistenceBitmap(DataOutputStream out, boolean[] bitmap) throws IOException {
for (int i = 0; i < 128; i++) {
byte b = 0;
for (int j = 0; j < 8; j++) {
if (bitmap[i * 8 + j]) {
b |= (1 << (7 - j));
}
}
out.writeByte(b);
}
}
private byte[] toByteArray(InputStream in) throws IOException {
ByteArrayOutputStream out = new ByteArrayOutputStream();
byte[] tempBuffer = new byte[4096];
int length;
while ((length = in.read(tempBuffer)) >= 0) {
out.write(tempBuffer, 0, length);
}
return out.toByteArray();
}
private static int getChunkIndex(int x, int z) {
return (x & 31) + ((z & 31) << 5);
}
private static int getTimestamp() {
return (int) (System.currentTimeMillis() / 1000L);
}
public void setOversized(int x, int z, boolean something) { }
public CompoundTag getOversizedData(int x, int z) throws IOException {
throw new IOException("getOversizedData is a stub " + this.regionFile);
}
public boolean isOversized(int x, int z) {
return false;
}
private class ChunkBuffer extends ByteArrayOutputStream {
private final ChunkPos pos;
public ChunkBuffer(ChunkPos chunkcoordintpair) {
super();
this.pos = chunkcoordintpair;
}
public void close() throws IOException {
ByteBuffer bytebuffer = ByteBuffer.wrap(this.buf, 0, this.count);
LinearRegionFile.this.write(this.pos, bytebuffer);
}
}
}