370 lines
10 KiB
TypeScript
370 lines
10 KiB
TypeScript
import sha1 from "sha1";
|
|
import { FileDef } from "../types/fileDef";
|
|
import fs from "fs";
|
|
import buildConfig from "../buildConfig";
|
|
import upath from "upath";
|
|
import requestretry from "requestretry";
|
|
import http from "http";
|
|
import { compareBufferToHashDef } from "./hashes";
|
|
import { execSync } from "child_process";
|
|
import { ModpackManifest, ModpackManifestFile, ExternalDependency } from "../types/modpackManifest";
|
|
import { fetchProject, fetchProjectsBulk } from "./curseForgeAPI";
|
|
import Bluebird from "bluebird";
|
|
import { VersionManifest } from "../types/versionManifest";
|
|
import { VersionsManifest } from "../types/versionsManifest";
|
|
import request from "requestretry";
|
|
import log from "fancy-log";
|
|
|
|
const LIBRARY_REG = /^(.+?):(.+?):(.+?)$/;
|
|
|
|
/**
|
|
* Parses the library name into path following the standard package naming convention.
|
|
*
|
|
* Turns `package:name:version` into `package/name/version/name-version`.
|
|
*/
|
|
export const libraryToPath = (library: string): string => {
|
|
const parsedLibrary = LIBRARY_REG.exec(library);
|
|
if (parsedLibrary) {
|
|
const pkg = parsedLibrary[1].replace(/\./g, "/");
|
|
const name = parsedLibrary[2];
|
|
const version = parsedLibrary[3];
|
|
|
|
const newURL = `${pkg}/${name}/${version}/${name}-${version}`;
|
|
|
|
return newURL;
|
|
}
|
|
};
|
|
|
|
/**
|
|
* Checks if given environmental variables are set. Throws otherwise.
|
|
*/
|
|
export const checkEnvironmentalVariables = (vars: string[]): void => {
|
|
vars.forEach((vari) => {
|
|
if (!process.env[vari] || process.env[vari] == "") {
|
|
throw new Error(`Environmental variable ${vari} is unset.`);
|
|
}
|
|
});
|
|
};
|
|
|
|
export enum RetrievedFileDefReason {
|
|
Downloaded,
|
|
CacheHit,
|
|
}
|
|
|
|
export interface RetrievedFileDef {
|
|
reason: RetrievedFileDefReason;
|
|
cachePath: string;
|
|
}
|
|
|
|
/**
|
|
* Downloads/fetches files from the Interwebs.
|
|
*
|
|
* Internally hashes the URL of the provided FileDef and looks it up in the cache directory.
|
|
* In case of no cache hit, downloads the file and stores within the cache directory for later use.
|
|
*/
|
|
export async function downloadOrRetrieveFileDef(fileDef: FileDef): Promise<RetrievedFileDef> {
|
|
const fileNameSha = sha1(fileDef.url);
|
|
|
|
const cachedFilePath = upath.join(buildConfig.downloaderCacheDirectory, fileNameSha);
|
|
if (fs.existsSync(cachedFilePath)) {
|
|
const file = await fs.promises.readFile(cachedFilePath);
|
|
|
|
if (file.length !== 0) {
|
|
const rFileDef = {
|
|
reason: RetrievedFileDefReason.CacheHit,
|
|
cachePath: cachedFilePath,
|
|
};
|
|
|
|
// Check hashes.
|
|
if (fileDef.hashes) {
|
|
if (
|
|
fileDef.hashes.every((hashDef) => {
|
|
return compareBufferToHashDef(file, hashDef);
|
|
})
|
|
) {
|
|
return rFileDef;
|
|
}
|
|
} else {
|
|
return rFileDef;
|
|
}
|
|
}
|
|
}
|
|
|
|
if (!fs.existsSync(buildConfig.downloaderCacheDirectory)) {
|
|
await fs.promises.mkdir(buildConfig.downloaderCacheDirectory, { recursive: true });
|
|
}
|
|
|
|
let handle: fs.promises.FileHandle;
|
|
try {
|
|
handle = await fs.promises.open(cachedFilePath, "w");
|
|
|
|
let hashFailed = false;
|
|
const retryStrategy = (err: Error, response: http.IncomingMessage, body: unknown) => {
|
|
// Verify hashes.
|
|
if (!err && fileDef.hashes && body) {
|
|
const success = fileDef.hashes.every((hashDef) => {
|
|
return compareBufferToHashDef(body as Buffer, hashDef);
|
|
});
|
|
|
|
if (!success) {
|
|
if (hashFailed) {
|
|
throw new Error(`Couldn't verify checksums of ${upath.basename(fileDef.url)}`);
|
|
}
|
|
|
|
hashFailed = true;
|
|
return true;
|
|
}
|
|
}
|
|
return requestretry.RetryStrategies.HTTPOrNetworkError(err, response, body);
|
|
};
|
|
|
|
const data: Buffer = Buffer.from(
|
|
await requestretry({
|
|
url: fileDef.url,
|
|
fullResponse: false,
|
|
encoding: null,
|
|
retryStrategy: retryStrategy,
|
|
maxAttempts: 5,
|
|
}),
|
|
);
|
|
|
|
await handle.write(data);
|
|
await handle.close();
|
|
|
|
return {
|
|
reason: RetrievedFileDefReason.Downloaded,
|
|
cachePath: cachedFilePath,
|
|
};
|
|
} catch (err) {
|
|
if (handle && (await handle.stat()).isFile()) {
|
|
log(`Couldn't download ${upath.basename(fileDef.url)}, cleaning up ${fileNameSha}...`);
|
|
|
|
await handle.close();
|
|
await fs.promises.unlink(cachedFilePath);
|
|
}
|
|
|
|
throw err;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Returns artifact name body depending on environment variables.
|
|
* Mostly intended to be called by CI/CD.
|
|
*/
|
|
export function makeArtifactNameBody(baseName: string): string {
|
|
// If the tag is provided by CI, simply just glue it to the base name.
|
|
if (process.env.GITHUB_TAG) {
|
|
return `${baseName}-${process.env.GITHUB_TAG}`;
|
|
}
|
|
// RC.
|
|
else if (process.env.RC_VERSION) {
|
|
return `${baseName}-${process.env.RC_VERSION.replace(/^v/, "")}`;
|
|
}
|
|
// If SHA is provided and the build isn't tagged, append both the branch and short SHA.
|
|
else if (process.env.GITHUB_SHA && process.env.GITHUB_REF && process.env.GITHUB_REF.startsWith("refs/heads/")) {
|
|
const shortCommit = process.env.GITHUB_SHA.substr(0, 7);
|
|
const branch = /refs\/heads\/(.+)/.exec(process.env.GITHUB_REF);
|
|
return `${baseName}-${branch[1]}-${shortCommit}`;
|
|
} else {
|
|
return baseName;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Fetches the last tag known to Git using the current branch.
|
|
* @param {string | nil} before Tag to get the tag before.
|
|
* @returns string Git tag.
|
|
* @throws
|
|
*/
|
|
export function getLastGitTag(before?: string): string {
|
|
if (before) {
|
|
before = `"${before}^"`;
|
|
}
|
|
|
|
return execSync(`git describe --abbrev=0 --tags ${before || ""}`)
|
|
.toString()
|
|
.trim();
|
|
}
|
|
|
|
/**
|
|
* Generates a changelog based on the two provided Git refs.
|
|
* @param {string} since Lower boundary Git ref.
|
|
* @param {string} to Upper boundary Git ref.
|
|
* @param {string[]} dirs Optional scopes.
|
|
*/
|
|
export function getChangeLog(since = "HEAD", to = "HEAD", dirs: string[] = undefined): string {
|
|
const command = [
|
|
"git log",
|
|
"--no-merges",
|
|
'--date="format:%d %b %Y"',
|
|
'--pretty="* %s - **%an** (%ad)"',
|
|
`${since}..${to}`,
|
|
];
|
|
|
|
if (dirs) {
|
|
command.push("--", dirs.join(" -- "));
|
|
}
|
|
|
|
return execSync(command.join(" ")).toString().trim();
|
|
}
|
|
|
|
/**
|
|
* Generates a changelog based on the two provided Git refs.
|
|
* @param {string} since Lower boundary Git ref.
|
|
* @param {string} to Upper boundary Git ref.
|
|
* @param {string[]} dirs Optional scopes.
|
|
*/
|
|
export function getFileAtRevision(path: string, revision = "HEAD"): string {
|
|
return execSync(`git show ${revision}:"${path}"`).toString().trim();
|
|
}
|
|
|
|
export interface ManifestFileListComparisonResult {
|
|
removed: string[];
|
|
modified: string[];
|
|
added: string[];
|
|
}
|
|
|
|
export async function compareAndExpandManifestDependencies(
|
|
oldFiles: ModpackManifest,
|
|
newFiles: ModpackManifest,
|
|
): Promise<ManifestFileListComparisonResult> {
|
|
// Map inputs for efficient joining.
|
|
const oldFileMap: { [key: number]: ModpackManifestFile } = oldFiles.files.reduce(
|
|
(map, file) => ((map[file.projectID] = file), map),
|
|
{},
|
|
);
|
|
const newFileMap: { [key: number]: ModpackManifestFile } = newFiles.files.reduce(
|
|
(map, file) => ((map[file.projectID] = file), map),
|
|
{},
|
|
);
|
|
|
|
const removed: string[] = [],
|
|
modified: string[] = [],
|
|
added: string[] = [];
|
|
|
|
// Create a distinct map of project IDs.
|
|
const projectIDs = Array.from(
|
|
new Set([...oldFiles.files.map((f) => f.projectID), ...newFiles.files.map((f) => f.projectID)]),
|
|
);
|
|
|
|
// Fetch projects in bulk and discard the result.
|
|
// Future calls to fetchProject() and fetchProjectsBulk() will hit the cache.
|
|
await fetchProjectsBulk(projectIDs);
|
|
|
|
await Bluebird.map(
|
|
projectIDs,
|
|
async (projectID) => {
|
|
const oldFileInfo = oldFileMap[projectID];
|
|
const newFileInfo = newFileMap[projectID];
|
|
|
|
// Doesn't exist in new, but exists in old. Removed. Left outer join.
|
|
if (!newFileInfo && oldFileInfo) {
|
|
removed.push((await fetchProject(oldFileInfo.projectID)).name);
|
|
}
|
|
// Doesn't exist in old, but exists in new. Added. Right outer join.
|
|
else if (newFileMap[projectID] && !oldFileMap[projectID]) {
|
|
added.push((await fetchProject(newFileInfo.projectID)).name);
|
|
}
|
|
// Exists in both. Modified? Inner join.
|
|
else if (oldFileInfo.fileID != newFileInfo.fileID) {
|
|
modified.push((await fetchProject(newFileInfo.projectID)).name);
|
|
}
|
|
},
|
|
{ concurrency: buildConfig.downloaderConcurrency },
|
|
);
|
|
|
|
// Compare external dependencies the same way.
|
|
const oldExternalMap: { [key: string]: ExternalDependency } = (oldFiles.externalDependencies || []).reduce(
|
|
(map, file) => ((map[file.name] = file), map),
|
|
{},
|
|
);
|
|
const newExternalMap: { [key: string]: ExternalDependency } = (newFiles.externalDependencies || []).reduce(
|
|
(map, file) => ((map[file.name] = file), map),
|
|
{},
|
|
);
|
|
|
|
const externalNames = Array.from(
|
|
new Set([
|
|
...(oldFiles.externalDependencies || []).map((dep) => dep.name),
|
|
...(newFiles.externalDependencies || []).map((dep) => dep.name),
|
|
]),
|
|
);
|
|
|
|
externalNames.forEach(async (name) => {
|
|
const oldDep = oldExternalMap[name];
|
|
const newDep = newExternalMap[name];
|
|
|
|
// Doesn't exist in new, but exists in old. Removed. Left outer join.
|
|
if (!newDep && oldDep) {
|
|
removed.push(oldDep.name);
|
|
}
|
|
// Doesn't exist in old, but exists in new. Added. Right outer join.
|
|
else if (newDep && !oldDep) {
|
|
added.push(newDep.name);
|
|
}
|
|
// Exists in both. Modified? Inner join.
|
|
else if (oldDep.url != newDep.url || oldDep.name != newDep.name) {
|
|
modified.push(newDep.name);
|
|
}
|
|
});
|
|
|
|
return {
|
|
removed: removed,
|
|
modified: modified,
|
|
added: added,
|
|
};
|
|
}
|
|
|
|
const LAUNCHERMETA_VERSION_MANIFEST = "https://launchermeta.mojang.com/mc/game/version_manifest.json";
|
|
|
|
/**
|
|
* Fetches the version manifest associated with the provided Minecraft version.
|
|
*
|
|
* @param minecraftVersion Minecraft version. (e. g., "1.12.2")
|
|
*/
|
|
export async function getVersionManifest(minecraftVersion: string): Promise<VersionManifest> {
|
|
/**
|
|
* Fetch the manifest file of all Minecraft versions.
|
|
*/
|
|
const manifest: VersionsManifest = await request({
|
|
uri: LAUNCHERMETA_VERSION_MANIFEST,
|
|
json: true,
|
|
fullResponse: false,
|
|
maxAttempts: 5,
|
|
});
|
|
|
|
const version = manifest.versions.find((x) => x.id == minecraftVersion);
|
|
if (!version) {
|
|
return null;
|
|
}
|
|
|
|
/**
|
|
* Fetch the version manifest file.
|
|
*/
|
|
const versionManifest: VersionManifest = await request({
|
|
uri: version.url,
|
|
json: true,
|
|
fullResponse: false,
|
|
maxAttempts: 5,
|
|
});
|
|
|
|
return versionManifest;
|
|
}
|
|
|
|
/**
|
|
* Returns a relative posix path from the first argument to the second.
|
|
*/
|
|
export function relative(from: string, to: string): string {
|
|
const broken = [from.split(upath.sep), to.split(upath.sep)];
|
|
|
|
while (broken.every((x) => x.length > 0) && broken[0][0] == broken[1][0]) {
|
|
broken.forEach((x) => x.shift());
|
|
}
|
|
|
|
if (broken.some((x) => x.length === 0)) {
|
|
throw new Error("Paths are not relative.");
|
|
}
|
|
|
|
return upath.join(...Array(broken[0].length - 1).fill(".."), ...broken[1]);
|
|
}
|