geforkt von Mirrors/Paper
8c5b837e05
Firstly, the old methods all routed to the CompletableFuture method. However, the CF method could not guarantee that if the caller was off-main that the future would be "completed" on-main. Since the callback methods used the CF one, this meant that the callback methods did not guarantee that the callbacks were to be called on the main thread. Now, all methods route to getChunkAtAsync(x, z, gen, urgent, cb) so that the methods with the callback are guaranteed to invoke the callback on the main thread. The CF behavior remains unchanged; it may still appear to complete on main if invoked off-main. Secondly, remove the scheduleOnMain invocation in the async chunk completion. This unnecessarily delays the callback by 1 tick. Thirdly, add getChunksAtAsync(minX, minZ, maxX, maxZ, ...) which will load chunks within an area. This method is provided as a helper as keeping all chunks loaded within an area can be complicated to implement for plugins (due to the lacking ticket API), and is already implemented internally anyways. Fourthly, remove the ticket addition that occured with getChunkAt and getChunkAtAsync. The ticket addition may delay the unloading of the chunk unnecessarily. It also fixes a very rare timing bug where the future/callback would be completed after the chunk unloads.
402 Zeilen
20 KiB
Diff
402 Zeilen
20 KiB
Diff
From 0000000000000000000000000000000000000000 Mon Sep 17 00:00:00 2001
|
|
From: Riley Park <rileysebastianpark@gmail.com>
|
|
Date: Tue, 16 Jul 2024 14:55:23 -0700
|
|
Subject: [PATCH] Bundle spark
|
|
|
|
|
|
diff --git a/build.gradle.kts b/build.gradle.kts
|
|
index 9e6c2a4630ce75e4115f76b5e7a1e0b50e8b3197..ee5f662ed0e84be997807a9faf97191ef4fc0449 100644
|
|
--- a/build.gradle.kts
|
|
+++ b/build.gradle.kts
|
|
@@ -76,6 +76,10 @@ dependencies {
|
|
implementation("io.papermc:reflection-rewriter-runtime:$reflectionRewriterVersion")
|
|
implementation("io.papermc:reflection-rewriter-proxy-generator:$reflectionRewriterVersion")
|
|
// Paper end - Remap reflection
|
|
+ // Paper start - spark
|
|
+ implementation("me.lucko:spark-api:0.1-20240720.200737-2")
|
|
+ implementation("me.lucko:spark-paper:1.10.105-SNAPSHOT")
|
|
+ // Paper end - spark
|
|
}
|
|
|
|
paperweight {
|
|
diff --git a/src/main/java/io/papermc/paper/SparksFly.java b/src/main/java/io/papermc/paper/SparksFly.java
|
|
new file mode 100644
|
|
index 0000000000000000000000000000000000000000..62e2d5704c348955bc8284dc2d54c933b7bcdd06
|
|
--- /dev/null
|
|
+++ b/src/main/java/io/papermc/paper/SparksFly.java
|
|
@@ -0,0 +1,211 @@
|
|
+package io.papermc.paper;
|
|
+
|
|
+import io.papermc.paper.configuration.GlobalConfiguration;
|
|
+import io.papermc.paper.plugin.entrypoint.classloader.group.PaperPluginClassLoaderStorage;
|
|
+import io.papermc.paper.plugin.provider.classloader.ConfiguredPluginClassLoader;
|
|
+import io.papermc.paper.plugin.provider.classloader.PaperClassLoaderStorage;
|
|
+import io.papermc.paper.util.MCUtil;
|
|
+import java.util.Collection;
|
|
+import java.util.List;
|
|
+import java.util.concurrent.ConcurrentLinkedQueue;
|
|
+import java.util.logging.Level;
|
|
+import java.util.logging.Logger;
|
|
+import me.lucko.spark.paper.api.Compatibility;
|
|
+import me.lucko.spark.paper.api.PaperClassLookup;
|
|
+import me.lucko.spark.paper.api.PaperScheduler;
|
|
+import me.lucko.spark.paper.api.PaperSparkModule;
|
|
+import net.kyori.adventure.text.Component;
|
|
+import net.kyori.adventure.text.format.TextColor;
|
|
+import net.minecraft.util.ExceptionCollector;
|
|
+import org.bukkit.Server;
|
|
+import org.bukkit.command.Command;
|
|
+import org.bukkit.command.CommandSender;
|
|
+import org.bukkit.craftbukkit.CraftServer;
|
|
+
|
|
+// It's like electricity.
|
|
+public final class SparksFly {
|
|
+ public static final String ID = "spark";
|
|
+ public static final String COMMAND_NAME = "spark";
|
|
+
|
|
+ private static final String PREFER_SPARK_PLUGIN_PROPERTY = "paper.preferSparkPlugin";
|
|
+
|
|
+ private static final int SPARK_YELLOW = 0xffc93a;
|
|
+
|
|
+ private final Logger logger;
|
|
+ private final PaperSparkModule spark;
|
|
+ private final ConcurrentLinkedQueue<Runnable> mainThreadTaskQueue;
|
|
+
|
|
+ private boolean enabled;
|
|
+ private boolean disabledInConfigurationWarningLogged;
|
|
+
|
|
+ public SparksFly(final Server server) {
|
|
+ this.mainThreadTaskQueue = new ConcurrentLinkedQueue<>();
|
|
+ this.logger = Logger.getLogger(ID);
|
|
+ this.logger.log(Level.INFO, "This server bundles the spark profiler. For more information please visit https://docs.papermc.io/paper/profiling");
|
|
+ this.spark = PaperSparkModule.create(Compatibility.VERSION_1_0, server, this.logger, new PaperScheduler() {
|
|
+ @Override
|
|
+ public void executeAsync(final Runnable runnable) {
|
|
+ MCUtil.scheduleAsyncTask(this.catching(runnable, "asynchronous"));
|
|
+ }
|
|
+
|
|
+ @Override
|
|
+ public void executeSync(final Runnable runnable) {
|
|
+ SparksFly.this.mainThreadTaskQueue.offer(this.catching(runnable, "synchronous"));
|
|
+ }
|
|
+
|
|
+ private Runnable catching(final Runnable runnable, final String type) {
|
|
+ return () -> {
|
|
+ try {
|
|
+ runnable.run();
|
|
+ } catch (final Throwable t) {
|
|
+ SparksFly.this.logger.log(Level.SEVERE, "An exception was encountered while executing a " + type + " spark task", t);
|
|
+ }
|
|
+ };
|
|
+ }
|
|
+ }, new PaperClassLookup() {
|
|
+ @Override
|
|
+ public Class<?> lookup(final String className) throws Exception {
|
|
+ final ExceptionCollector<ClassNotFoundException> exceptions = new ExceptionCollector<>();
|
|
+ try {
|
|
+ return Class.forName(className);
|
|
+ } catch (final ClassNotFoundException e) {
|
|
+ exceptions.add(e);
|
|
+ for (final ConfiguredPluginClassLoader loader : ((PaperPluginClassLoaderStorage) PaperClassLoaderStorage.instance()).getGlobalGroup().getClassLoaders()) {
|
|
+ try {
|
|
+ final Class<?> loadedClass = loader.loadClass(className, true, false, true);
|
|
+ if (loadedClass != null) {
|
|
+ return loadedClass;
|
|
+ }
|
|
+ } catch (final ClassNotFoundException exception) {
|
|
+ exceptions.add(exception);
|
|
+ }
|
|
+ }
|
|
+ exceptions.throwIfPresent();
|
|
+ return null;
|
|
+ }
|
|
+ }
|
|
+ });
|
|
+ }
|
|
+
|
|
+ public void executeMainThreadTasks() {
|
|
+ Runnable task;
|
|
+ while ((task = this.mainThreadTaskQueue.poll()) != null) {
|
|
+ task.run();
|
|
+ }
|
|
+ }
|
|
+
|
|
+ public void enableEarlyIfRequested() {
|
|
+ if (!isPluginPreferred() && shouldEnableImmediately()) {
|
|
+ this.enable();
|
|
+ }
|
|
+ }
|
|
+
|
|
+ public void enableBeforePlugins() {
|
|
+ if (!isPluginPreferred()) {
|
|
+ this.enable();
|
|
+ }
|
|
+ }
|
|
+
|
|
+ public void enableAfterPlugins(final Server server) {
|
|
+ final boolean isPluginPreferred = isPluginPreferred();
|
|
+ final boolean isPluginEnabled = isPluginEnabled(server);
|
|
+ if (!isPluginPreferred || !isPluginEnabled) {
|
|
+ if (isPluginPreferred && !this.enabled) {
|
|
+ this.logger.log(Level.INFO, "The spark plugin has been preferred but was not loaded. The bundled spark profiler will enabled instead.");
|
|
+ }
|
|
+ this.enable();
|
|
+ }
|
|
+ }
|
|
+
|
|
+ private void enable() {
|
|
+ if (!this.enabled) {
|
|
+ if (GlobalConfiguration.get().spark.enabled) {
|
|
+ this.enabled = true;
|
|
+ this.spark.enable();
|
|
+ } else {
|
|
+ if (!this.disabledInConfigurationWarningLogged) {
|
|
+ this.logger.log(Level.INFO, "The spark profiler will not be enabled because it is currently disabled in the configuration.");
|
|
+ this.disabledInConfigurationWarningLogged = true;
|
|
+ }
|
|
+ }
|
|
+ }
|
|
+ }
|
|
+
|
|
+ public void disable() {
|
|
+ if (this.enabled) {
|
|
+ this.spark.disable();
|
|
+ this.enabled = false;
|
|
+ }
|
|
+ }
|
|
+
|
|
+ public void registerCommandBeforePlugins(final Server server) {
|
|
+ if (!isPluginPreferred()) {
|
|
+ this.registerCommand(server);
|
|
+ }
|
|
+ }
|
|
+
|
|
+ public void registerCommandAfterPlugins(final Server server) {
|
|
+ if ((!isPluginPreferred() || !isPluginEnabled(server)) && server.getCommandMap().getCommand(COMMAND_NAME) == null) {
|
|
+ this.registerCommand(server);
|
|
+ }
|
|
+ }
|
|
+
|
|
+ private void registerCommand(final Server server) {
|
|
+ server.getCommandMap().register(COMMAND_NAME, "paper", new CommandImpl(COMMAND_NAME, this.spark.getPermissions()));
|
|
+ }
|
|
+
|
|
+ public void tickStart() {
|
|
+ this.spark.onServerTickStart();
|
|
+ }
|
|
+
|
|
+ public void tickEnd(final double duration) {
|
|
+ this.spark.onServerTickEnd(duration);
|
|
+ }
|
|
+
|
|
+ void executeCommand(final CommandSender sender, final String[] args) {
|
|
+ this.spark.executeCommand(sender, args);
|
|
+ }
|
|
+
|
|
+ List<String> tabComplete(final CommandSender sender, final String[] args) {
|
|
+ return this.spark.tabComplete(sender, args);
|
|
+ }
|
|
+
|
|
+ public static boolean isPluginPreferred() {
|
|
+ return Boolean.getBoolean(PREFER_SPARK_PLUGIN_PROPERTY);
|
|
+ }
|
|
+
|
|
+ private static boolean isPluginEnabled(final Server server) {
|
|
+ return server.getPluginManager().isPluginEnabled(ID);
|
|
+ }
|
|
+
|
|
+ private static boolean shouldEnableImmediately() {
|
|
+ return GlobalConfiguration.get().spark.enableImmediately;
|
|
+ }
|
|
+
|
|
+ public static final class CommandImpl extends Command {
|
|
+ CommandImpl(final String name, final Collection<String> permissions) {
|
|
+ super(name);
|
|
+ this.setPermission(String.join(";", permissions));
|
|
+ }
|
|
+
|
|
+ @Override
|
|
+ public boolean execute(final CommandSender sender, final String commandLabel, final String[] args) {
|
|
+ final SparksFly spark = ((CraftServer) sender.getServer()).spark;
|
|
+ if (spark.enabled) {
|
|
+ spark.executeCommand(sender, args);
|
|
+ } else {
|
|
+ sender.sendMessage(Component.text("The spark profiler is currently disabled.", TextColor.color(SPARK_YELLOW)));
|
|
+ }
|
|
+ return true;
|
|
+ }
|
|
+
|
|
+ @Override
|
|
+ public List<String> tabComplete(final CommandSender sender, final String alias, final String[] args) throws IllegalArgumentException {
|
|
+ final SparksFly spark = ((CraftServer) sender.getServer()).spark;
|
|
+ if (spark.enabled) {
|
|
+ return spark.tabComplete(sender, args);
|
|
+ }
|
|
+ return List.of();
|
|
+ }
|
|
+ }
|
|
+}
|
|
diff --git a/src/main/java/io/papermc/paper/plugin/provider/source/FileProviderSource.java b/src/main/java/io/papermc/paper/plugin/provider/source/FileProviderSource.java
|
|
index 6b8ed8a0baaf4a57d20e57cec3400af5561ddd79..48604e7f96adc9e226e034054c5e2bad0b024eb5 100644
|
|
--- a/src/main/java/io/papermc/paper/plugin/provider/source/FileProviderSource.java
|
|
+++ b/src/main/java/io/papermc/paper/plugin/provider/source/FileProviderSource.java
|
|
@@ -1,6 +1,9 @@
|
|
package io.papermc.paper.plugin.provider.source;
|
|
|
|
+import com.mojang.logging.LogUtils;
|
|
+import io.papermc.paper.SparksFly;
|
|
import io.papermc.paper.plugin.PluginInitializerManager;
|
|
+import io.papermc.paper.plugin.configuration.PluginMeta;
|
|
import io.papermc.paper.plugin.entrypoint.EntrypointHandler;
|
|
import io.papermc.paper.plugin.provider.type.PluginFileType;
|
|
import org.bukkit.plugin.InvalidPluginException;
|
|
@@ -17,12 +20,14 @@ import java.nio.file.attribute.BasicFileAttributes;
|
|
import java.util.Set;
|
|
import java.util.function.Function;
|
|
import java.util.jar.JarFile;
|
|
+import org.slf4j.Logger;
|
|
|
|
/**
|
|
* Loads a plugin provider at the given plugin jar file path.
|
|
*/
|
|
public class FileProviderSource implements ProviderSource<Path, Path> {
|
|
|
|
+ private static final Logger LOGGER = LogUtils.getClassLogger();
|
|
private final Function<Path, String> contextChecker;
|
|
private final boolean applyRemap;
|
|
|
|
@@ -82,6 +87,12 @@ public class FileProviderSource implements ProviderSource<Path, Path> {
|
|
);
|
|
}
|
|
|
|
+ final PluginMeta config = type.getConfig(file);
|
|
+ if ((config.getName().equals("spark") && config.getMainClass().equals("me.lucko.spark.bukkit.BukkitSparkPlugin")) && !SparksFly.isPluginPreferred()) {
|
|
+ LOGGER.info("The spark plugin will not be loaded as this server bundles the spark profiler.");
|
|
+ return;
|
|
+ }
|
|
+
|
|
type.register(entrypointHandler, file, context);
|
|
}
|
|
|
|
diff --git a/src/main/java/net/minecraft/server/MinecraftServer.java b/src/main/java/net/minecraft/server/MinecraftServer.java
|
|
index b29dbab3c26c078eb993183d808149d958857f3a..38b5eef0b5ff38f9a8dfa59822d7b41336108f49 100644
|
|
--- a/src/main/java/net/minecraft/server/MinecraftServer.java
|
|
+++ b/src/main/java/net/minecraft/server/MinecraftServer.java
|
|
@@ -764,6 +764,8 @@ public abstract class MinecraftServer extends ReentrantBlockableEventLoop<TickTa
|
|
// Paper end - Configurable player collision
|
|
|
|
this.server.enablePlugins(org.bukkit.plugin.PluginLoadOrder.POSTWORLD);
|
|
+ this.server.spark.registerCommandBeforePlugins(this.server); // Paper - spark
|
|
+ this.server.spark.enableAfterPlugins(this.server); // Paper - spark
|
|
if (io.papermc.paper.plugin.PluginInitializerManager.instance().pluginRemapper != null) io.papermc.paper.plugin.PluginInitializerManager.instance().pluginRemapper.pluginsEnabled(); // Paper - Remap plugins
|
|
io.papermc.paper.command.brigadier.PaperCommands.INSTANCE.setValid(); // Paper - reset invalid state for event fire below
|
|
io.papermc.paper.plugin.lifecycle.event.LifecycleEventRunner.INSTANCE.callReloadableRegistrarEvent(io.papermc.paper.plugin.lifecycle.event.types.LifecycleEvents.COMMANDS, io.papermc.paper.command.brigadier.PaperCommands.INSTANCE, org.bukkit.plugin.Plugin.class, io.papermc.paper.plugin.lifecycle.event.registrar.ReloadableRegistrarEvent.Cause.INITIAL); // Paper - call commands event for regular plugins
|
|
@@ -1051,6 +1053,7 @@ public abstract class MinecraftServer extends ReentrantBlockableEventLoop<TickTa
|
|
Commands.COMMAND_SENDING_POOL.shutdownNow(); // Paper - Perf: Async command map building; Shutdown and don't bother finishing
|
|
// CraftBukkit start
|
|
if (this.server != null) {
|
|
+ this.server.spark.disable(); // Paper - spark
|
|
this.server.disablePlugins();
|
|
this.server.waitForAsyncTasksShutdown(); // Paper - Wait for Async Tasks during shutdown
|
|
}
|
|
@@ -1244,6 +1247,7 @@ public abstract class MinecraftServer extends ReentrantBlockableEventLoop<TickTa
|
|
// tasks are default scheduled at -1 + delay, and first tick will tick at 1
|
|
final long actualDoneTimeMs = System.currentTimeMillis() - org.bukkit.craftbukkit.Main.BOOT_TIME.toEpochMilli(); // Paper - Add total time
|
|
LOGGER.info("Done ({})! For help, type \"help\"", String.format(java.util.Locale.ROOT, "%.3fs", actualDoneTimeMs / 1000.00D)); // Paper - Add total time
|
|
+ this.server.spark.enableBeforePlugins(); // Paper - spark
|
|
org.spigotmc.WatchdogThread.tick();
|
|
// Paper end - Improved Watchdog Support
|
|
org.spigotmc.WatchdogThread.hasStarted = true; // Paper
|
|
@@ -1626,17 +1630,21 @@ public abstract class MinecraftServer extends ReentrantBlockableEventLoop<TickTa
|
|
}
|
|
|
|
if (this.emptyTicks >= j) {
|
|
+ this.server.spark.tickStart(); // Paper - spark
|
|
if (this.emptyTicks == j) {
|
|
MinecraftServer.LOGGER.info("Server empty for {} seconds, pausing", this.pauseWhileEmptySeconds());
|
|
this.autoSave();
|
|
}
|
|
|
|
this.server.getScheduler().mainThreadHeartbeat(); // CraftBukkit
|
|
+ this.server.spark.executeMainThreadTasks(); // Paper - spark
|
|
this.tickConnection();
|
|
+ this.server.spark.tickEnd(((double)(System.nanoTime() - lastTick) / 1000000D)); // Paper - spark
|
|
return;
|
|
}
|
|
}
|
|
|
|
+ this.server.spark.tickStart(); // Paper - spark
|
|
new com.destroystokyo.paper.event.server.ServerTickStartEvent(this.tickCount+1).callEvent(); // Paper - Server Tick Events
|
|
++this.tickCount;
|
|
this.tickRateManager.tick();
|
|
@@ -1654,11 +1662,13 @@ public abstract class MinecraftServer extends ReentrantBlockableEventLoop<TickTa
|
|
ProfilerFiller gameprofilerfiller = Profiler.get();
|
|
|
|
this.runAllTasks(); // Paper - move runAllTasks() into full server tick (previously for timings)
|
|
+ this.server.spark.executeMainThreadTasks(); // Paper - spark
|
|
// Paper start - Server Tick Events
|
|
long endTime = System.nanoTime();
|
|
long remaining = (TICK_TIME - (endTime - lastTick)) - catchupTime;
|
|
new com.destroystokyo.paper.event.server.ServerTickEndEvent(this.tickCount, ((double)(endTime - lastTick) / 1000000D), remaining).callEvent();
|
|
// Paper end - Server Tick Events
|
|
+ this.server.spark.tickEnd(((double)(endTime - lastTick) / 1000000D)); // Paper - spark
|
|
gameprofilerfiller.push("tallying");
|
|
long k = Util.getNanos() - i;
|
|
int l = this.tickCount % 100;
|
|
diff --git a/src/main/java/net/minecraft/server/dedicated/DedicatedServer.java b/src/main/java/net/minecraft/server/dedicated/DedicatedServer.java
|
|
index c06863578c5d654706d93e73059d89c12ae502a5..17a158ff6ce6520b69a5a0032ba4c05449dd0cf8 100644
|
|
--- a/src/main/java/net/minecraft/server/dedicated/DedicatedServer.java
|
|
+++ b/src/main/java/net/minecraft/server/dedicated/DedicatedServer.java
|
|
@@ -223,6 +223,7 @@ public class DedicatedServer extends MinecraftServer implements ServerInterface
|
|
this.paperConfigurations.initializeGlobalConfiguration(this.registryAccess());
|
|
this.paperConfigurations.initializeWorldDefaultsConfiguration(this.registryAccess());
|
|
// Paper end - initialize global and world-defaults configuration
|
|
+ this.server.spark.enableEarlyIfRequested(); // Paper - spark
|
|
// Paper start - fix converting txt to json file; convert old users earlier after PlayerList creation but before file load/save
|
|
if (this.convertOldUsers()) {
|
|
this.getProfileCache().save(false); // Paper
|
|
@@ -232,6 +233,7 @@ public class DedicatedServer extends MinecraftServer implements ServerInterface
|
|
org.spigotmc.WatchdogThread.doStart(org.spigotmc.SpigotConfig.timeoutTime, org.spigotmc.SpigotConfig.restartOnCrash); // Paper - start watchdog thread
|
|
thread.start(); // Paper - Enhance console tab completions for brigadier commands; start console thread after MinecraftServer.console & PaperConfig are initialized
|
|
io.papermc.paper.command.PaperCommands.registerCommands(this); // Paper - setup /paper command
|
|
+ this.server.spark.registerCommandBeforePlugins(this.server); // Paper - spark
|
|
com.destroystokyo.paper.Metrics.PaperMetrics.startMetrics(); // Paper - start metrics
|
|
com.destroystokyo.paper.VersionHistoryManager.INSTANCE.getClass(); // Paper - load version history now
|
|
|
|
diff --git a/src/main/java/org/bukkit/craftbukkit/CraftServer.java b/src/main/java/org/bukkit/craftbukkit/CraftServer.java
|
|
index 6323738fe96c8e815a0835e6d8b03125c36ecf90..806e56cb60235a99f468d36a059fdbd54c2d46e3 100644
|
|
--- a/src/main/java/org/bukkit/craftbukkit/CraftServer.java
|
|
+++ b/src/main/java/org/bukkit/craftbukkit/CraftServer.java
|
|
@@ -312,6 +312,7 @@ public final class CraftServer implements Server {
|
|
public static Exception excessiveVelEx; // Paper - Velocity warnings
|
|
private final io.papermc.paper.logging.SysoutCatcher sysoutCatcher = new io.papermc.paper.logging.SysoutCatcher(); // Paper
|
|
private final io.papermc.paper.potion.PaperPotionBrewer potionBrewer; // Paper - Custom Potion Mixes
|
|
+ public final io.papermc.paper.SparksFly spark; // Paper - spark
|
|
|
|
// Paper start - Folia region threading API
|
|
private final io.papermc.paper.threadedregions.scheduler.FallbackRegionScheduler regionizedScheduler = new io.papermc.paper.threadedregions.scheduler.FallbackRegionScheduler();
|
|
@@ -477,6 +478,7 @@ public final class CraftServer implements Server {
|
|
}
|
|
this.potionBrewer = new io.papermc.paper.potion.PaperPotionBrewer(console); // Paper - custom potion mixes
|
|
datapackManager = new io.papermc.paper.datapack.PaperDatapackManager(console.getPackRepository()); // Paper
|
|
+ this.spark = new io.papermc.paper.SparksFly(this); // Paper - spark
|
|
}
|
|
|
|
public boolean getCommandBlockOverride(String command) {
|
|
@@ -1103,6 +1105,7 @@ public final class CraftServer implements Server {
|
|
this.reloadData();
|
|
org.spigotmc.SpigotConfig.registerCommands(); // Spigot
|
|
io.papermc.paper.command.PaperCommands.registerCommands(this.console); // Paper
|
|
+ this.spark.registerCommandBeforePlugins(this); // Paper - spark
|
|
this.overrideAllCommandBlockCommands = this.commandsConfiguration.getStringList("command-block-overrides").contains("*");
|
|
this.ignoreVanillaPermissions = this.commandsConfiguration.getBoolean("ignore-vanilla-permissions");
|
|
|
|
@@ -1131,6 +1134,7 @@ public final class CraftServer implements Server {
|
|
this.loadPlugins();
|
|
this.enablePlugins(PluginLoadOrder.STARTUP);
|
|
this.enablePlugins(PluginLoadOrder.POSTWORLD);
|
|
+ this.spark.registerCommandAfterPlugins(this); // Paper - spark
|
|
if (io.papermc.paper.plugin.PluginInitializerManager.instance().pluginRemapper != null) io.papermc.paper.plugin.PluginInitializerManager.instance().pluginRemapper.pluginsEnabled(); // Paper - Remap plugins
|
|
// Paper start - brigadier command API
|
|
io.papermc.paper.command.brigadier.PaperCommands.INSTANCE.setValid(); // to clear invalid state for event fire below
|