/* * This file is part of fabric-loom, licensed under the MIT License (MIT). * * Copyright (c) 2016-2025 FabricMC * * Permission is hereby granted, free of charge, to any person obtaining a copy * of this software and associated documentation files (the "Software"), to deal * in the Software without restriction, including without limitation the rights * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell * copies of the Software, and to permit persons to whom the Software is * furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice shall be included in all * copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. */ package net.fabricmc.loom.task; import java.io.BufferedReader; import java.io.BufferedWriter; import java.io.File; import java.io.IOException; import java.io.UncheckedIOException; import java.lang.reflect.Constructor; import java.lang.reflect.InvocationTargetException; import java.nio.charset.StandardCharsets; import java.nio.file.Files; import java.nio.file.Path; import java.nio.file.StandardCopyOption; import java.time.Duration; import java.util.Collection; import java.util.HashMap; import java.util.Iterator; import java.util.List; import java.util.Locale; import java.util.Map; import java.util.Objects; import java.util.StringJoiner; import java.util.UUID; import java.util.stream.Collectors; import java.util.stream.Stream; import javax.inject.Inject; import dev.architectury.loom.forge.ForgeSourcesService; import org.gradle.api.file.ConfigurableFileCollection; import org.gradle.api.file.FileCollection; import org.gradle.api.file.RegularFileProperty; import org.gradle.api.provider.Property; import org.gradle.api.services.ServiceReference; import org.gradle.api.tasks.Input; import org.gradle.api.tasks.InputFiles; import org.gradle.api.tasks.Internal; import org.gradle.api.tasks.Nested; import org.gradle.api.tasks.Optional; import org.gradle.api.tasks.OutputFile; import org.gradle.api.tasks.TaskAction; import org.gradle.api.tasks.UntrackedTask; import org.gradle.api.tasks.options.Option; import org.gradle.internal.logging.progress.ProgressLoggerFactory; import org.gradle.process.ExecOperations; import org.gradle.workers.WorkAction; import org.gradle.workers.WorkParameters; import org.gradle.workers.WorkQueue; import org.gradle.workers.WorkerExecutor; import org.gradle.workers.internal.WorkerDaemonClientsManager; import org.jetbrains.annotations.ApiStatus; import org.jetbrains.annotations.Nullable; import net.fabricmc.loom.api.decompilers.DecompilationMetadata; import net.fabricmc.loom.api.decompilers.DecompilerOptions; import net.fabricmc.loom.api.decompilers.LoomDecompiler; import net.fabricmc.loom.configuration.providers.minecraft.MinecraftJar; import net.fabricmc.loom.configuration.providers.minecraft.mapped.AbstractMappedMinecraftProvider; import net.fabricmc.loom.decompilers.ClassLineNumbers; import net.fabricmc.loom.decompilers.LineNumberRemapper; import net.fabricmc.loom.decompilers.cache.CachedData; import net.fabricmc.loom.decompilers.cache.CachedFileStoreImpl; import net.fabricmc.loom.decompilers.cache.CachedJarProcessor; import net.fabricmc.loom.task.service.SourceMappingsService; import net.fabricmc.loom.task.service.UnpickService; import net.fabricmc.loom.util.Checksum; import net.fabricmc.loom.util.Constants; import net.fabricmc.loom.util.ExceptionUtil; import net.fabricmc.loom.util.FileSystemUtil; import net.fabricmc.loom.util.IOStringConsumer; import net.fabricmc.loom.util.Platform; import net.fabricmc.loom.util.gradle.GradleUtils; import net.fabricmc.loom.util.gradle.SyncTaskBuildService; import net.fabricmc.loom.util.gradle.ThreadedProgressLoggerConsumer; import net.fabricmc.loom.util.gradle.ThreadedSimpleProgressLogger; import net.fabricmc.loom.util.gradle.WorkerDaemonClientsManagerHelper; import net.fabricmc.loom.util.gradle.daemon.DaemonUtils; import net.fabricmc.loom.util.ipc.IPCClient; import net.fabricmc.loom.util.ipc.IPCServer; import net.fabricmc.loom.util.service.ScopedServiceFactory; import net.fabricmc.loom.util.service.ServiceFactory; import net.fabricmc.mappingio.tree.MemoryMappingTree; @UntrackedTask(because = "Manually invoked, has internal caching") public abstract class GenerateSourcesTask extends AbstractLoomTask { private static final String CACHE_VERSION = "v1"; private final DecompilerOptions decompilerOptions; /** * The jar name to decompile, {@link MinecraftJar#getName()}. */ @Input public abstract Property getInputJarName(); @InputFiles // Only contains a single file protected abstract ConfigurableFileCollection getClassesInputJar(); @InputFiles protected abstract ConfigurableFileCollection getClasspath(); @InputFiles protected abstract ConfigurableFileCollection getMinecraftCompileLibraries(); @OutputFile public abstract RegularFileProperty getSourcesOutputJar(); // Contains the remapped linenumbers @OutputFile protected abstract ConfigurableFileCollection getClassesOutputJar(); // Single jar @Input @Option(option = "use-cache", description = "Use the decompile cache") @ApiStatus.Experimental public abstract Property getUseCache(); @Input @Option(option = "reset-cache", description = "When set the cache will be reset") @ApiStatus.Experimental public abstract Property getResetCache(); // Internal inputs @ApiStatus.Internal @Nested protected abstract Property getMappings(); // Internal outputs @ApiStatus.Internal @Internal protected abstract RegularFileProperty getDecompileCacheFile(); @ApiStatus.Internal @Input protected abstract Property getMaxCachedFiles(); @ApiStatus.Internal @Input protected abstract Property getMaxCacheFileAge(); // Injects @Inject protected abstract WorkerExecutor getWorkerExecutor(); @Inject protected abstract ExecOperations getExecOperations(); @Inject protected abstract WorkerDaemonClientsManager getWorkerDaemonClientsManager(); @Inject protected abstract ProgressLoggerFactory getProgressLoggerFactory(); @Nested protected abstract Property getDaemonUtilsContext(); @Nested @Optional protected abstract Property getUnpickOptions(); @Nested @Optional protected abstract Property getForgeSourcesOptions(); // Prevent Gradle from running two gen sources tasks in parallel @ServiceReference(SyncTaskBuildService.NAME) abstract Property getSyncTask(); @Inject public GenerateSourcesTask(DecompilerOptions decompilerOptions) { this.decompilerOptions = decompilerOptions; getClassesInputJar().setFrom(getInputJarName().map(minecraftJarName -> { final List minecraftJars = getExtension().getNamedMinecraftProvider().getMinecraftJars(); for (MinecraftJar minecraftJar : minecraftJars) { if (minecraftJar.getName().equals(minecraftJarName)) { final Path backupJarPath = AbstractMappedMinecraftProvider.getBackupJarPath(minecraftJar); if (Files.notExists(backupJarPath)) { throw new IllegalStateException("Input minecraft jar not found at: " + backupJarPath); } return backupJarPath.toFile(); } } throw new IllegalStateException("Input minecraft jar not found: " + getInputJarName().get()); })); getClassesOutputJar().setFrom(getInputJarName().map(minecraftJarName -> { final List minecraftJars = getExtension().getNamedMinecraftProvider().getMinecraftJars(); for (MinecraftJar minecraftJar : minecraftJars) { if (minecraftJar.getName().equals(minecraftJarName)) { return minecraftJar.toFile(); } } throw new IllegalStateException("Input minecraft jar not found: " + getInputJarName().get()); })); getClasspath().from(decompilerOptions.getClasspath()).finalizeValueOnRead(); dependsOn(decompilerOptions.getClasspath().getBuiltBy()); getMinecraftCompileLibraries().from(getProject().getConfigurations().getByName(Constants.Configurations.MINECRAFT_COMPILE_LIBRARIES)); getDecompileCacheFile().set(getExtension().getFiles().getDecompileCache(CACHE_VERSION)); getUseCache().convention(true); getResetCache().convention(getExtension().refreshDeps()); getMappings().set(SourceMappingsService.create(getProject())); getMaxCachedFiles().set(GradleUtils.getIntegerPropertyProvider(getProject(), Constants.Properties.DECOMPILE_CACHE_MAX_FILES).orElse(50_000)); getMaxCacheFileAge().set(GradleUtils.getIntegerPropertyProvider(getProject(), Constants.Properties.DECOMPILE_CACHE_MAX_AGE).orElse(90)); getDaemonUtilsContext().set(getProject().getObjects().newInstance(DaemonUtils.Context.class, getProject())); getUnpickOptions().set(UnpickService.createOptions(this)); getForgeSourcesOptions().set(ForgeSourcesService.createOptions(getProject())); mustRunAfter(getProject().getTasks().withType(AbstractRemapJarTask.class)); } @TaskAction public void run() throws IOException { final Platform platform = Platform.CURRENT; if (!platform.getArchitecture().is64Bit()) { throw new UnsupportedOperationException("GenSources task requires a 64bit JVM to run due to the memory requirements."); } try (ScopedServiceFactory serviceFactory = new ScopedServiceFactory()) { if (!getUseCache().get()) { getLogger().info("Not using decompile cache."); try (var timer = new Timer("Decompiled sources")) { runWithoutCache(serviceFactory); } catch (Exception e) { ExceptionUtil.processException(e, getDaemonUtilsContext().get()); throw ExceptionUtil.createDescriptiveWrapper(RuntimeException::new, "Failed to decompile", e); } return; } getLogger().info("Using decompile cache."); try (var timer = new Timer("Decompiled sources with cache")) { final Path cacheFile = getDecompileCacheFile().getAsFile().get().toPath(); if (getResetCache().get()) { getLogger().warn("Resetting decompile cache"); Files.deleteIfExists(cacheFile); } // TODO ensure we have a lock on this file to prevent multiple tasks from running at the same time Files.createDirectories(cacheFile.getParent()); if (Files.exists(cacheFile)) { try (FileSystemUtil.Delegate fs = FileSystemUtil.getJarFileSystem(cacheFile, true)) { // Success, cache exists and can be read } catch (IOException e) { getLogger().warn("Discarding invalid decompile cache file: {}", cacheFile, e); Files.delete(cacheFile); } } try (FileSystemUtil.Delegate fs = FileSystemUtil.getJarFileSystem(cacheFile, true)) { runWithCache(serviceFactory, fs.getRoot()); } } catch (Exception e) { ExceptionUtil.processException(e, getDaemonUtilsContext().get()); throw ExceptionUtil.createDescriptiveWrapper(RuntimeException::new, "Failed to decompile", e); } } } private void runWithCache(ServiceFactory serviceFactory, Path cacheRoot) throws IOException { final Path classesInputJar = getClassesInputJar().getSingleFile().toPath(); final Path sourcesOutputJar = getSourcesOutputJar().get().getAsFile().toPath(); final Path classesOutputJar = getClassesOutputJar().getSingleFile().toPath(); final var cacheRules = new CachedFileStoreImpl.CacheRules(getMaxCachedFiles().get(), Duration.ofDays(getMaxCacheFileAge().get())); final var decompileCache = new CachedFileStoreImpl<>(cacheRoot, CachedData.SERIALIZER, cacheRules); final String cacheKey = getCacheKey(serviceFactory); final CachedJarProcessor cachedJarProcessor = new CachedJarProcessor(decompileCache, cacheKey); final CachedJarProcessor.WorkRequest workRequest; getLogger().info("Decompile cache key: {}", cacheKey); getLogger().debug("Decompile cache rules: {}", cacheRules); try (var timer = new Timer("Prepare job")) { workRequest = cachedJarProcessor.prepareJob(classesInputJar); } final CachedJarProcessor.WorkJob job = workRequest.job(); final CachedJarProcessor.CacheStats cacheStats = workRequest.stats(); getLogger().lifecycle("Decompile cache stats: {} hits, {} misses", cacheStats.hits(), cacheStats.misses()); ClassLineNumbers outputLineNumbers = null; if (job instanceof CachedJarProcessor.WorkToDoJob workToDoJob) { Path workInputJar = workToDoJob.incomplete(); @Nullable Path existingClasses = (job instanceof CachedJarProcessor.PartialWorkJob partialWorkJob) ? partialWorkJob.existingClasses() : null; if (usingUnpick()) { try (var timer = new Timer("Unpick")) { UnpickService unpick = serviceFactory.get(getUnpickOptions()); workInputJar = unpick.unpickJar(workInputJar, existingClasses); } } try (var timer = new Timer("Decompile")) { outputLineNumbers = runDecompileJob(workInputJar, workToDoJob.output(), existingClasses); removeForgeInnerClassSources(workToDoJob.output()); outputLineNumbers = filterForgeLineNumbers(outputLineNumbers); } if (Files.notExists(workToDoJob.output())) { throw new RuntimeException("Failed to decompile sources"); } } else if (job instanceof CachedJarProcessor.CompletedWorkJob completedWorkJob) { // Nothing to do :) } // The final output sources jar Files.deleteIfExists(sourcesOutputJar); try (var timer = new Timer("Complete job")) { cachedJarProcessor.completeJob(sourcesOutputJar, job, outputLineNumbers); } getLogger().info("Decompiled sources written to {}", sourcesOutputJar); // Remap the line numbers with the new and existing numbers final ClassLineNumbers existingLinenumbers = workRequest.lineNumbers(); final ClassLineNumbers lineNumbers = ClassLineNumbers.merge(existingLinenumbers, outputLineNumbers); applyLineNumbers(lineNumbers, classesInputJar, classesOutputJar); try (var timer = new Timer("Prune cache")) { decompileCache.prune(); } } private void runWithoutCache(ServiceFactory serviceFactory) throws IOException { final Path classesInputJar = getClassesInputJar().getSingleFile().toPath(); final Path sourcesOutputJar = getSourcesOutputJar().get().getAsFile().toPath(); final Path classesOutputJar = getClassesOutputJar().getSingleFile().toPath(); Path workClassesJar = classesInputJar; if (usingUnpick()) { try (var timer = new Timer("Unpick")) { UnpickService unpick = serviceFactory.get(getUnpickOptions()); workClassesJar = unpick.unpickJar(workClassesJar, null); } } ClassLineNumbers lineNumbers; try (var timer = new Timer("Decompile")) { lineNumbers = runDecompileJob(workClassesJar, sourcesOutputJar, null); removeForgeInnerClassSources(sourcesOutputJar); lineNumbers = filterForgeLineNumbers(lineNumbers); } if (Files.notExists(sourcesOutputJar)) { throw new RuntimeException("Failed to decompile sources"); } getLogger().info("Decompiled sources written to {}", sourcesOutputJar); applyLineNumbers(lineNumbers, classesInputJar, classesOutputJar); } private void applyLineNumbers(@Nullable ClassLineNumbers lineNumbers, Path classesInputJar, Path classesOutputJar) throws IOException { if (lineNumbers == null) { getLogger().info("No line numbers to remap, skipping remapping"); return; } final Path tempJar = Files.createTempFile("loom", "linenumber-remap.jar"); Files.delete(tempJar); try (var timer = new Timer("Remap line numbers")) { remapLineNumbers(lineNumbers, classesInputJar, tempJar); } Files.move(tempJar, classesOutputJar, StandardCopyOption.REPLACE_EXISTING); } private String getCacheKey(ServiceFactory serviceFactory) { var sj = new StringJoiner(","); sj.add(getDecompilerCheckKey()); if (usingUnpick()) { UnpickService unpick = serviceFactory.get(getUnpickOptions()); sj.add(unpick.getUnpickCacheKey()); } SourceMappingsService mappingsService = serviceFactory.get(getMappings()); String mappingsHash = mappingsService.getProcessorHash(); if (mappingsHash != null) { sj.add(mappingsHash); } getLogger().info("Decompile cache data: {}", sj); return Checksum.of(sj.toString()).sha256().hex(); } private String getDecompilerCheckKey() { var sj = new StringJoiner(","); sj.add(decompilerOptions.getDecompilerClassName().get()); sj.add(Checksum.of(decompilerOptions.getClasspath()).sha256().hex()); for (Map.Entry entry : decompilerOptions.getOptions().get().entrySet()) { sj.add(entry.getKey() + "=" + entry.getValue()); } return sj.toString(); } @Nullable private ClassLineNumbers runDecompileJob(Path inputJar, Path outputJar, @Nullable Path existingJar) throws IOException { final Platform platform = Platform.CURRENT; final Path lineMapFile = File.createTempFile("loom", "linemap").toPath(); Files.delete(lineMapFile); if (!platform.supportsUnixDomainSockets()) { getLogger().warn("Decompile worker logging disabled as Unix Domain Sockets is not supported on your operating system."); doWork(null, inputJar, outputJar, lineMapFile, existingJar); // Inject Forge's own sources try (var serviceFactory = new ScopedServiceFactory()) { final @Nullable ForgeSourcesService service = serviceFactory.getOrNull(getForgeSourcesOptions()); if (service != null) { service.addForgeSources(inputJar, outputJar); } } return readLineNumbers(lineMapFile); } // Set up the IPC path to get the log output back from the forked JVM final Path ipcPath = Files.createTempFile("loom", "ipc"); Files.deleteIfExists(ipcPath); try (ThreadedProgressLoggerConsumer loggerConsumer = new ThreadedProgressLoggerConsumer(getLogger(), getProgressLoggerFactory(), decompilerOptions.getName(), "Decompiling minecraft sources"); IPCServer logReceiver = new IPCServer(ipcPath, loggerConsumer)) { doWork(logReceiver, inputJar, outputJar, lineMapFile, existingJar); } catch (InterruptedException e) { throw new RuntimeException("Failed to shutdown log receiver", e); } finally { Files.deleteIfExists(ipcPath); } // Inject Forge's own sources try (var serviceFactory = new ScopedServiceFactory()) { final @Nullable ForgeSourcesService service = serviceFactory.getOrNull(getForgeSourcesOptions()); if (service != null) { service.addForgeSources(inputJar, outputJar); } } return readLineNumbers(lineMapFile); } @Nullable private ClassLineNumbers filterForgeLineNumbers(@Nullable ClassLineNumbers lineNumbers) { if (lineNumbers == null) { return null; } if (getModPlatform().get().isForgeLike()) { // Remove Forge and NeoForge classes from linemap // TODO: We should instead not decompile Forge's classes at all var lineMap = new HashMap(); for (Map.Entry entry : lineNumbers.lineMap().entrySet()) { String name = entry.getKey(); if (!name.startsWith("net/minecraftforge/") && !name.startsWith("net/neoforged/")) { lineMap.put(name, entry.getValue()); } } return new ClassLineNumbers(lineMap); } else { return lineNumbers; } } /** * Some inner classes orders are messed up with forge recompilation, I don't know if that is why the decompiler * would occasionally split out extra inner classes (where with normal fabric setups it doesn't happen), * but this is a workaround for that. */ private void removeForgeInnerClassSources(Path sourcesJar) throws IOException { if (!getModPlatform().get().isForgeLike()) return; try (FileSystemUtil.Delegate outputFs = FileSystemUtil.getJarFileSystem(sourcesJar, false); Stream walk = Files.walk(outputFs.getRoot())) { Iterator iterator = walk.iterator(); while (iterator.hasNext()) { final Path fsPath = iterator.next(); if (fsPath.startsWith("/META-INF/")) { continue; } if (!Files.isRegularFile(fsPath)) { continue; } if (fsPath.toString().substring(outputFs.getRoot().toString().length()).indexOf('$') != -1) { Files.delete(fsPath); } } } } private void remapLineNumbers(ClassLineNumbers lineNumbers, Path inputJar, Path outputJar) throws IOException { Objects.requireNonNull(lineNumbers, "lineNumbers"); final var remapper = new LineNumberRemapper(lineNumbers); remapper.process(inputJar, outputJar); final Path lineMap = inputJar.resolveSibling(inputJar.getFileName() + ".linemap.txt"); try (BufferedWriter writer = Files.newBufferedWriter(lineMap)) { lineNumbers.write(writer); } getLogger().info("Wrote linemap to {}", lineMap); } private void doWork(@Nullable IPCServer ipcServer, Path inputJar, Path outputJar, Path linemapFile, @Nullable Path existingClasses) { final String jvmMarkerValue = UUID.randomUUID().toString(); final WorkQueue workQueue = createWorkQueue(jvmMarkerValue); workQueue.submit(DecompileAction.class, params -> { params.getDecompilerOptions().set(decompilerOptions.toDto()); params.getInputJar().set(inputJar.toFile()); params.getOutputJar().set(outputJar.toFile()); params.getLinemapFile().set(linemapFile.toFile()); params.getMappings().set(getMappings()); if (ipcServer != null) { params.getIPCPath().set(ipcServer.getPath().toFile()); } params.getClassPath().setFrom(getMinecraftCompileLibraries()); if (existingClasses != null) { params.getClassPath().from(existingClasses); } // Architectury params.getForge().set(getModPlatform().get().isForgeLike()); }); try { workQueue.await(); } finally { if (ipcServer != null) { boolean stopped = WorkerDaemonClientsManagerHelper.stopIdleJVM(getWorkerDaemonClientsManager(), jvmMarkerValue); if (!stopped && ipcServer.hasReceivedMessage()) { getLogger().info("Failed to stop decompile worker JVM, it may have already been stopped?"); } } } } private WorkQueue createWorkQueue(String jvmMarkerValue) { if (!useProcessIsolation()) { return getWorkerExecutor().classLoaderIsolation(spec -> { spec.getClasspath().from(getClasspath()); }); } return getWorkerExecutor().processIsolation(spec -> { spec.forkOptions(forkOptions -> { forkOptions.setMinHeapSize(String.format(Locale.ENGLISH, "%dm", Math.min(512, decompilerOptions.getMemory().get()))); forkOptions.setMaxHeapSize(String.format(Locale.ENGLISH, "%dm", decompilerOptions.getMemory().get())); forkOptions.systemProperty(WorkerDaemonClientsManagerHelper.MARKER_PROP, jvmMarkerValue); }); spec.getClasspath().from(getClasspath()); }); } private boolean useProcessIsolation() { // Useful if you want to debug the decompiler, make sure you run gradle with enough memory. return !Boolean.getBoolean("fabric.loom.genSources.debug"); } private boolean usingUnpick() { return getUnpickOptions().isPresent(); } public interface DecompileParams extends WorkParameters { Property getDecompilerOptions(); RegularFileProperty getInputJar(); RegularFileProperty getOutputJar(); RegularFileProperty getLinemapFile(); Property getMappings(); RegularFileProperty getIPCPath(); ConfigurableFileCollection getClassPath(); // Architectury Property getForge(); } public abstract static class DecompileAction implements WorkAction { @Override public void execute() { if (!getParameters().getIPCPath().isPresent() || !Platform.CURRENT.supportsUnixDomainSockets()) { // Does not support unix domain sockets, print to sout. doDecompile(System.out::println); return; } final Path ipcPath = getParameters().getIPCPath().get().getAsFile().toPath(); try (IPCClient ipcClient = new IPCClient(ipcPath)) { doDecompile(new ThreadedSimpleProgressLogger(ipcClient)); } catch (Exception e) { throw ExceptionUtil.createDescriptiveWrapper(RuntimeException::new, "Failed to decompile", e); } } private void doDecompile(IOStringConsumer logger) { final Path inputJar = getParameters().getInputJar().get().getAsFile().toPath(); final Path linemap = getParameters().getLinemapFile().get().getAsFile().toPath(); final Path outputJar = getParameters().getOutputJar().get().getAsFile().toPath(); final DecompilerOptions.Dto decompilerOptions = getParameters().getDecompilerOptions().get(); final LoomDecompiler decompiler; try { final String className = decompilerOptions.className(); final Constructor decompilerConstructor = getDecompilerConstructor(className); Objects.requireNonNull(decompilerConstructor, "%s must have a no args constructor".formatted(className)); decompiler = decompilerConstructor.newInstance(); } catch (InstantiationException | IllegalAccessException | InvocationTargetException e) { throw new RuntimeException("Failed to create decompiler", e); } try (var serviceFactory = new ScopedServiceFactory()) { final SourceMappingsService mappingsService = serviceFactory.get(getParameters().getMappings()); final var metadata = new DecompilationMetadata( decompilerOptions.maxThreads(), mappingsService.getMappingsFile(), getLibraries(), logger, decompilerOptions.options() ); decompiler.decompile( inputJar, outputJar, linemap, metadata ); // Close the decompile loggers try { metadata.logger().accept(ThreadedProgressLoggerConsumer.CLOSE_LOGGERS); } catch (IOException e) { throw new UncheckedIOException("Failed to close loggers", e); } } catch (IOException e) { throw new UncheckedIOException(e); } } private Collection getLibraries() { return toPaths(getParameters().getClassPath()); } static Collection toPaths(FileCollection files) { return files.getFiles().stream().map(File::toPath).collect(Collectors.toSet()); } } public static File getJarFileWithSuffix(String suffix, Path runtimeJar) { final String path = runtimeJar.toFile().getAbsolutePath(); if (!path.toLowerCase(Locale.ROOT).endsWith(".jar")) { throw new RuntimeException("Invalid mapped JAR path: " + path); } return new File(path.substring(0, path.length() - 4) + suffix); } static File getJarFileWithSuffix(RegularFileProperty runtimeJar, String suffix) { return getJarFileWithSuffix(suffix, runtimeJar.get().getAsFile().toPath()); } @Nullable private static ClassLineNumbers readLineNumbers(Path linemapFile) throws IOException { if (Files.notExists(linemapFile)) { return null; } try (BufferedReader reader = Files.newBufferedReader(linemapFile, StandardCharsets.UTF_8)) { return ClassLineNumbers.readMappings(reader); } catch (Exception e) { throw new IOException("Failed to read line number map: " + linemapFile, e); } } private static Constructor getDecompilerConstructor(String clazz) { try { //noinspection unchecked return (Constructor) Class.forName(clazz).getConstructor(); } catch (NoSuchMethodException e) { return null; } catch (ClassNotFoundException e) { throw new RuntimeException(e); } } public interface MappingsProcessor { boolean transform(MemoryMappingTree mappings); } private final class Timer implements AutoCloseable { private final String name; private final long start; Timer(String name) { this.name = name; this.start = System.currentTimeMillis(); } @Override public void close() { getLogger().info("{} took {}ms", name, System.currentTimeMillis() - start); } } }