name: Release on: workflow_dispatch: inputs: ref: description: 'Ref to build (for Pull Requests, use refs/pull/NNN/head)' required: true # repository_dispatch: # # client_payload should be the same as the inputs for workflow_dispatch. # types: # - Build* # release: # types: # - published jobs: source: name: Source runs-on: ubuntu-20.04 outputs: version: ${{ steps.metadata.outputs.version }} is_tag: ${{ steps.metadata.outputs.is_tag }} trigger_type: ${{ steps.metadata.outputs.trigger_type }} folder: ${{ steps.metadata.outputs.folder }} steps: - name: Checkout (Release) if: github.event_name == 'release' uses: actions/checkout@v3 with: # We generate a changelog; for this we need the full git log. fetch-depth: 0 - name: Checkout (Manual) if: github.event_name == 'workflow_dispatch' uses: actions/checkout@v3 with: ref: ${{ github.event.inputs.ref }} # We generate a changelog; for this we need the full git log. fetch-depth: 0 - name: Checkout (Trigger) if: github.event_name == 'repository_dispatch' uses: actions/checkout@v3 with: ref: ${{ github.event.client_payload.ref }} # We generate a changelog; for this we need the full git log. fetch-depth: 0 - name: Check valid branch name run: | if [ "${{ github.event_name }}" = "workflow_dispatch" ]; then REF="${{ github.event.inputs.ref }}" elif [ "${{ github.event_name }}" = "repository_dispatch" ]; then REF="${{ github.event.client_payload.ref }}" else REF="${{ github.ref }}" fi # Check if we are a tag. if [ -n "$(git name-rev --name-only --tags --no-undefined HEAD 2>/dev/null || false)" ]; then exit 0 fi # Check if the checkout caused the branch to be named. if [ "$(git rev-parse --abbrev-ref HEAD)" != "HEAD" ]; then exit 0 fi # Check if this was a pull request. if [ -n "$(echo ${REF} | grep '^refs/pull/[0-9]*')" ]; then PULL=$(echo ${REF} | cut -d/ -f3) git checkout -b pr${PULL} fi # Are we still in a detached state? Error out. if [ "$(git rev-parse --abbrev-ref HEAD)" == "HEAD" ]; then echo "The 'ref' given resulted in a checkout of a detached HEAD." echo "We cannot detect the version for these checkout accurate." echo "" echo "If you want to build a Pull Request, make sure you use 'refs/pull/NNN/head'." echo "" echo "Cancelling build, as without a version we cannot store the artifacts." exit 1 fi - name: Generate metadata id: metadata run: | echo "::group::Prepare metadata files" cmake -DGENERATE_OTTDREV=1 -P cmake/scripts/FindVersion.cmake ./.github/changelog.sh > .changelog TZ='UTC' date +"%Y-%m-%d %H:%M UTC" > .release_date head -1 .ottdrev-vc | cut -f 1 -d$'\t' > .version if [ $(head -1 .ottdrev-vc | cut -f 5 -d$'\t') = '1' ]; then # Assume that all tags are always releases. Why else make a tag? IS_TAG="true" FOLDER="${{ env.FOLDER_RELEASES }}" TRIGGER_TYPE="new-tag" else IS_TAG="false" BRANCH=$(git symbolic-ref -q HEAD | sed 's@.*/@@') if [ -z "${BRANCH}" ]; then echo "Internal error: branch name is empty." echo "An earlier step should have prevented this from happening." echo "Cancelling build, as without a branch name we cannot store the artifacts" exit 1 fi if [ "${BRANCH}" = "${{ env.NIGHTLIES_BRANCH }}" ]; then # The "master" branch is special, and we call a nightly. FOLDER="${{ env.FOLDER_NIGHTLIES }}/$(date +%Y)" TRIGGER_TYPE="new-master" else # All other branches, which can be builds of Pull Requests, are # put in their own folder. FOLDER="${{ env.FOLDER_BRANCHES }}/${BRANCH}" TRIGGER_TYPE="new-branch" fi fi mkdir -p build/bundles cp .changelog build/bundles/changelog.txt cp .release_date build/bundles/released.txt cp README.md build/bundles/README.md echo "::endgroup::" echo "Release Date: $(cat .release_date)" echo "Revision: $(cat .ottdrev-vc)" echo "Version: $(cat .version)" echo "Is tag: ${IS_TAG}" echo "Folder on CDN: ${FOLDER}" echo "Workflow trigger: ${TRIGGER_TYPE}" echo "version=$(cat .version)" >> $GITHUB_OUTPUT echo "is_tag=${IS_TAG}" >> $GITHUB_OUTPUT echo "folder=${FOLDER}" >> $GITHUB_OUTPUT echo "trigger_type=${TRIGGER_TYPE}" >> $GITHUB_OUTPUT env: NIGHTLIES_BRANCH: master FOLDER_RELEASES: openttd-releases FOLDER_NIGHTLIES: openttd-nightlies FOLDER_BRANCHES: openttd-branches - name: Remove VCS information run: | rm -rf .git - name: Create bundles run: | FOLDER_NAME=openttd-${{ steps.metadata.outputs.version }} # Rename the folder to openttd-NNN mkdir ${FOLDER_NAME} find . -maxdepth 1 -not -name . -not -name build -not -name ${FOLDER_NAME} -exec mv {} ${FOLDER_NAME}/ \; echo "::group::Create tarball (xz) bundle" tar --xz -cvf build/bundles/${FOLDER_NAME}-source.tar.xz ${FOLDER_NAME} echo "::endgroup::" # This tarball is only to be used within this workflow. echo "::group::Create tarball (gz) bundle" tar --gzip -cvf source.tar.gz ${FOLDER_NAME} echo "::endgroup::" echo "::group::Create zip bundle" zip -9 -r build/bundles/${FOLDER_NAME}-source.zip ${FOLDER_NAME} echo "::endgroup::" - name: Store bundles uses: actions/upload-artifact@v3 with: name: openttd-source path: build/bundles/* retention-days: 5 - name: Store source (for other jobs) uses: actions/upload-artifact@v3 with: name: internal-source path: source.tar.gz retention-days: 1 docs: name: Docs needs: source runs-on: ubuntu-20.04 steps: - name: Download source uses: actions/download-artifact@v3 with: name: internal-source - name: Unpack source run: | tar -xf source.tar.gz --strip-components=1 - name: Install dependencies run: | echo "::group::Update apt" sudo apt-get update echo "::endgroup::" echo "::group::Install dependencies" sudo apt-get install -y --no-install-recommends \ doxygen \ # EOF echo "::endgroup::" env: DEBIAN_FRONTEND: noninteractive - name: Build run: | mkdir -p ${GITHUB_WORKSPACE}/build cd ${GITHUB_WORKSPACE}/build echo "::group::CMake" cmake ${GITHUB_WORKSPACE} \ -DCMAKE_BUILD_TYPE=RelWithDebInfo \ -DOPTION_DOCS_ONLY=ON \ # EOF echo "::endgroup::" echo "::group::Build" cmake --build . --target docs echo "::endgroup::" - name: Create bundles run: | BASENAME=openttd-${{ needs.source.outputs.version }} cd ${GITHUB_WORKSPACE}/build mv docs/source ${BASENAME}-docs mv docs/ai-api ${BASENAME}-docs-ai mv docs/gs-api ${BASENAME}-docs-gs mkdir -p bundles echo "::group::Create docs bundle" tar --xz -cf bundles/${BASENAME}-docs.tar.xz ${BASENAME}-docs echo "::endgroup::" echo "::group::Create AI API docs bundle" tar --xz -cf bundles/${BASENAME}-docs-ai.tar.xz ${BASENAME}-docs-ai echo "::endgroup::" echo "::group::Create GameScript API docs bundle" tar --xz -cf bundles/${BASENAME}-docs-gs.tar.xz ${BASENAME}-docs-gs echo "::endgroup::" - name: Store bundles uses: actions/upload-artifact@v3 with: name: openttd-docs path: build/bundles/*.tar.xz retention-days: 5 linux: name: Linux (Generic) needs: source runs-on: ubuntu-20.04 container: # manylinux2014 is based on CentOS 7, but already has a lot of things # installed and preconfigured. It makes it easier to build OpenTTD. image: quay.io/pypa/manylinux2014_x86_64 steps: - name: Download source uses: actions/download-artifact@v3 with: name: internal-source - name: Unpack source run: | tar -xf source.tar.gz --strip-components=1 - name: Install dependencies run: | echo "::group::Install dependencies" yum install -y \ fontconfig-devel \ freetype-devel \ libicu-devel \ libpng-devel \ libpng-devel \ lzo-devel \ SDL2-devel \ wget \ xz-devel \ libzstd-devel \ zlib-devel \ # EOF echo "::endgroup::" # The yum variant of fluidsynth depends on all possible audio drivers, # like jack, ALSA, pulseaudio, etc. This is not really useful for us, # as we route the output of fluidsynth back via our sound driver, and # as such do not use these audio driver outputs at all. So instead, # we compile fluidsynth ourselves, with as little dependencies as # possible. This currently means it picks up SDL2, but this is fine, # as we need SDL2 anyway. echo "::group::Install fluidsynth" wget https://github.com/FluidSynth/fluidsynth/archive/v2.1.6.tar.gz tar xf v2.1.6.tar.gz ( cd fluidsynth-2.1.6 mkdir build cd build cmake .. -DCMAKE_BUILD_TYPE=RelWithDebInfo -DCMAKE_INSTALL_PREFIX=/usr cmake --build . -j $(nproc) cmake --install . ) echo "::endgroup::" - name: Install GCC problem matcher uses: ammaraskar/gcc-problem-matcher@master - name: Build run: | mkdir -p build cd build echo "::group::CMake" cmake ${GITHUB_WORKSPACE} \ -DCMAKE_BUILD_TYPE=RelWithDebInfo \ -DOPTION_PACKAGE_DEPENDENCIES=ON \ # EOF echo "::endgroup::" echo "::group::Build" echo "Running on $(nproc) cores" cmake --build . -j $(nproc) echo "::endgroup::" - name: Create bundles run: | cd ${GITHUB_WORKSPACE}/build echo "::group::Run CPack" cpack echo "::endgroup::" echo "::group::Cleanup" # Remove the sha256 files CPack generates; we will do this ourself at # the end of this workflow. rm -f bundles/*.sha256 echo "::endgroup::" - name: Store bundles uses: actions/upload-artifact@v3 with: name: openttd-linux-generic path: build/bundles retention-days: 5 linux-distro: name: Linux (Distros) needs: source if: needs.source.outputs.is_tag == 'true' strategy: fail-fast: false matrix: include: - container_image: "ubuntu:18.04" bundle_name: "bionic" compiler: "g++-8" c_compiler: "gcc-8" - container_image: "ubuntu:20.04" bundle_name: "focal" compiler: "g++" c_compiler: "gcc" - container_image: "ubuntu:22.04" bundle_name: "jammy" compiler: "g++" c_compiler: "gcc" - container_image: "debian:buster" bundle_name: "buster" compiler: "g++" c_compiler: "gcc" - container_image: "debian:bullseye" bundle_name: "bullseye" compiler: "g++" c_compiler: "gcc" runs-on: ubuntu-20.04 container: image: ${{ matrix.container_image }} steps: - name: Download source uses: actions/download-artifact@v3 with: name: internal-source - name: Unpack source run: | tar -xf source.tar.gz --strip-components=1 - name: Install dependencies run: | echo "::group::Update apt" apt-get update echo "::endgroup::" echo "::group::Install dependencies" apt-get install -y --no-install-recommends \ cmake \ debhelper \ ${{ matrix.compiler }} \ ${{ matrix.c_compiler }} \ git \ make \ openssl \ libfontconfig-dev \ libfluidsynth-dev \ libicu-dev \ liblzma-dev \ libzstd-dev \ liblzo2-dev \ libsdl2-dev \ lsb-release \ zlib1g-dev \ # EOF echo "::endgroup::" env: DEBIAN_FRONTEND: noninteractive - name: Install GCC problem matcher uses: ammaraskar/gcc-problem-matcher@master - name: Build run: | mkdir -p build cd build echo "::group::CMake" CXX=${{ matrix.compiler }} CC=${{ matrix.c_compiler }} cmake ${GITHUB_WORKSPACE} \ -DCMAKE_BUILD_TYPE=RelWithDebInfo \ -DCMAKE_INSTALL_PREFIX=/usr \ # EOF echo "::endgroup::" echo "::group::Build" echo "Running on $(nproc) cores" # Ubuntu 18.04 cmake does not support -j so we pass the option to the native tool cmake --build . -- -j $(nproc) echo "::endgroup::" - name: Create bundles run: | cd ${GITHUB_WORKSPACE}/build echo "::group::Run CPack" cpack echo "::endgroup::" echo "::group::Cleanup" # Remove the sha256 files CPack generates; we will do this ourself at # the end of this workflow. rm -f bundles/*.sha256 echo "::endgroup::" - name: Store bundles uses: actions/upload-artifact@v3 with: name: openttd-linux-${{ matrix.bundle_name }} path: build/bundles retention-days: 5 macos: name: MacOS needs: source runs-on: macos-11 env: MACOSX_DEPLOYMENT_TARGET: 10.9 steps: - name: Download source uses: actions/download-artifact@v3 with: name: internal-source - name: Unpack source run: | tar -xf source.tar.gz --strip-components=1 - name: Install dependencies env: HOMEBREW_NO_AUTO_UPDATE: 1 HOMEBREW_NO_INSTALL_CLEANUP: 1 run: | brew install pandoc - name: Prepare cache key id: key run: | echo "image=$ImageOS-$ImageVersion" >> $GITHUB_OUTPUT - name: Enable vcpkg cache uses: actions/cache@v3 with: path: /usr/local/share/vcpkg/installed key: ${{ steps.key.outputs.image }}-vcpkg-release-0 # Increase the number whenever dependencies are modified restore-keys: | ${{ steps.key.outputs.image }}-vcpkg-release ${{ steps.key.outputs.image }}-vcpkg-x64 - name: Prepare vcpkg run: | vcpkg install \ liblzma:x64-osx \ liblzma:arm64-osx \ libpng:x64-osx \ libpng:arm64-osx \ lzo:x64-osx \ lzo:arm64-osx \ zlib:x64-osx \ zlib:arm64-osx \ zstd:x64-osx \ zstd:arm64-osx \ # EOF - name: Install GCC problem matcher uses: ammaraskar/gcc-problem-matcher@master - name: Build tools run: | mkdir build-host cd build-host echo "::group::CMake" cmake ${GITHUB_WORKSPACE} \ -DCMAKE_BUILD_TYPE=RelWithDebInfo \ -DOPTION_TOOLS_ONLY=ON \ # EOF echo "::endgroup::" echo "::group::Build tools" echo "Running on $(sysctl -n hw.logicalcpu) cores" cmake --build . -j $(sysctl -n hw.logicalcpu) --target tools echo "::endgroup::" - name: Import code signing certificates if: ${{ false }} # disable for now uses: Apple-Actions/import-codesign-certs@v1 with: # The certificates in a PKCS12 file encoded as a base64 string p12-file-base64: ${{ secrets.APPLE_DEVELOPER_CERTIFICATE_P12_BASE64 }} # The password used to import the PKCS12 file. p12-password: ${{ secrets.APPLE_DEVELOPER_CERTIFICATE_PASSWORD }} # If this is run on a fork, there may not be a certificate set up - continue in this case continue-on-error: true - name: Build arm64 run: | mkdir build-arm64 cd build-arm64 echo "::group::CMake" cmake ${GITHUB_WORKSPACE} \ -DCMAKE_OSX_ARCHITECTURES=arm64 \ -DVCPKG_TARGET_TRIPLET=arm64-osx \ -DCMAKE_TOOLCHAIN_FILE=/usr/local/share/vcpkg/scripts/buildsystems/vcpkg.cmake \ -DHOST_BINARY_DIR=${GITHUB_WORKSPACE}/build-host \ -DCMAKE_BUILD_TYPE=RelWithDebInfo \ # EOF echo "::endgroup::" echo "::group::Build" echo "Running on $(sysctl -n hw.logicalcpu) cores" cmake --build . -j $(sysctl -n hw.logicalcpu) echo "::endgroup::" - name: Build x64 run: | mkdir build-x64 cd build-x64 echo "::group::CMake" cmake ${GITHUB_WORKSPACE} \ -DCMAKE_OSX_ARCHITECTURES=x86_64 \ -DVCPKG_TARGET_TRIPLET=x64-osx \ -DCMAKE_TOOLCHAIN_FILE=/usr/local/share/vcpkg/scripts/buildsystems/vcpkg.cmake \ -DHOST_BINARY_DIR=${GITHUB_WORKSPACE}/build-host \ -DCMAKE_BUILD_TYPE=RelWithDebInfo \ -DCPACK_BUNDLE_APPLE_CERT_APP=${{ secrets.APPLE_DEVELOPER_CERTIFICATE_ID }} \ "-DCPACK_BUNDLE_APPLE_CODESIGN_PARAMETER=--deep -f --options runtime" \ -DAPPLE_UNIVERSAL_PACKAGE=1 \ # EOF echo "::endgroup::" echo "::group::Build" echo "Running on $(sysctl -n hw.logicalcpu) cores" cmake --build . -j $(sysctl -n hw.logicalcpu) echo "::endgroup::" - name: Create bundles run: | cd build-x64 echo "::group::Create universal binary" # Combine the `openttd` binaries from each build into a single file lipo -create -output openttd-universal ../build-*/openttd mv openttd-universal openttd echo "::endgroup::" echo "::group::Run CPack" cpack echo "::endgroup::" echo "::group::Cleanup" # Remove the sha256 files CPack generates; we will do this ourself at # the end of this workflow. rm -f bundles/*.sha256 echo "::endgroup::" - name: Install gon if: ${{ false }} # disable for now env: HOMEBREW_NO_AUTO_UPDATE: 1 HOMEBREW_NO_INSTALL_CLEANUP: 1 run: | brew tap mitchellh/gon brew install mitchellh/gon/gon - name: Notarize env: AC_USERNAME: ${{ secrets.APPLE_DEVELOPER_APP_USERNAME }} AC_PASSWORD: ${{ secrets.APPLE_DEVELOPER_APP_PASSWORD }} run: | cd build-x64 ../os/macosx/notarize.sh - name: Build zip run: | cd build-x64 pushd _CPack_Packages/*/Bundle/openttd-*/ # Remove the Applications symlink from the staging folder rm -f Applications # Remove the original dmg built by CPack to avoid a conflict when resolving # the zip_filename variable below rm -f ../*.dmg zip_filename=(../openttd-*) # Package up the existing, notarised .app into a zip file zip -r -9 ${zip_filename}.zip OpenTTD.app popd # Now move it into place to be uploaded mv _CPack_Packages/*/Bundle/openttd-*.zip bundles/ - name: Store bundles uses: actions/upload-artifact@v3 with: name: openttd-macos-universal path: build-x64/bundles retention-days: 5 windows: name: Windows needs: source strategy: fail-fast: false matrix: include: - arch: x86 host: x86 - arch: x64 host: x64 - arch: arm64 host: x64_arm64 runs-on: windows-latest steps: - name: Download source uses: actions/download-artifact@v3 with: name: internal-source - name: Unpack source shell: bash run: | tar -xf source.tar.gz --strip-components=1 - name: Install dependencies shell: bash run: | choco install pandoc - name: Prepare cache key id: key shell: powershell run: | # Work around caching failure with GNU tar New-Item -Type Junction -Path vcpkg -Target c:\vcpkg Write-Output "image=$env:ImageOS-$env:ImageVersion" >> $env:GITHUB_OUTPUT - name: Enable vcpkg cache uses: actions/cache@v3 with: path: vcpkg/installed key: ${{ steps.key.outputs.image }}-vcpkg-${{ matrix.arch }}-0 # Increase the number whenever dependencies are modified restore-keys: | ${{ steps.key.outputs.image }}-vcpkg-${{ matrix.arch }} - name: Prepare vcpkg shell: bash run: | vcpkg install --triplet=${{ matrix.arch }}-windows-static \ liblzma \ libpng \ lzo \ zlib \ # EOF - name: Install MSVC problem matcher uses: ammaraskar/msvc-problem-matcher@master - name: Configure developer command prompt for tools uses: ilammy/msvc-dev-cmd@v1 with: arch: x64 - name: Build tools shell: bash run: | mkdir build-host cd build-host echo "::group::CMake" cmake ${GITHUB_WORKSPACE} \ -GNinja \ -DOPTION_TOOLS_ONLY=ON \ -DCMAKE_BUILD_TYPE=RelWithDebInfo \ # EOF echo "::endgroup::" echo "::group::Build" cmake --build . --target tools echo "::endgroup::" - name: Configure developer command prompt for ${{ matrix.arch }} uses: ilammy/msvc-dev-cmd@v1 with: arch: ${{ matrix.host }} - name: Import code signing certificate if: ${{ false }} # Disabled shell: powershell # If this is run on a fork, there may not be a certificate set up - continue in this case continue-on-error: true run: | $tempFile = [System.IO.Path]::GetTempFileName() $bytes = [System.Convert]::FromBase64String($env:WINDOWS_CERTIFICATE_P12) [IO.File]::WriteAllBytes($tempFile, $bytes) $pwd = ConvertTo-SecureString $env:WINDOWS_CERTIFICATE_PASSWORD -AsPlainText -Force Import-PfxCertificate -FilePath $tempFile -CertStoreLocation Cert:\CurrentUser\My -Password $pwd Remove-Item $tempFile env: WINDOWS_CERTIFICATE_P12: ${{ secrets.WINDOWS_CERTIFICATE_P12 }} WINDOWS_CERTIFICATE_PASSWORD: ${{ secrets.WINDOWS_CERTIFICATE_PASSWORD }} - name: Build (with installer) if: needs.source.outputs.is_tag == 'true' shell: bash run: | mkdir build cd build echo "::group::CMake" cmake ${GITHUB_WORKSPACE} \ -GNinja \ -DVCPKG_TARGET_TRIPLET=${{ matrix.arch }}-windows-static \ -DCMAKE_TOOLCHAIN_FILE="c:\vcpkg\scripts\buildsystems\vcpkg.cmake" \ -DOPTION_USE_NSIS=ON \ -DHOST_BINARY_DIR=${GITHUB_WORKSPACE}/build-host \ -DCMAKE_BUILD_TYPE=RelWithDebInfo \ -DWINDOWS_CERTIFICATE_COMMON_NAME="${WINDOWS_CERTIFICATE_COMMON_NAME}" \ # EOF echo "::endgroup::" echo "::group::Build" cmake --build . echo "::endgroup::" env: WINDOWS_CERTIFICATE_COMMON_NAME: ${{ secrets.WINDOWS_CERTIFICATE_COMMON_NAME }} - name: Build (without installer) if: needs.source.outputs.is_tag != 'true' shell: bash run: | mkdir build cd build echo "::group::CMake" cmake ${GITHUB_WORKSPACE} \ -GNinja \ -DVCPKG_TARGET_TRIPLET=${{ matrix.arch }}-windows-static \ -DCMAKE_TOOLCHAIN_FILE="c:\vcpkg\scripts\buildsystems\vcpkg.cmake" \ -DHOST_BINARY_DIR=${GITHUB_WORKSPACE}/build-host \ -DCMAKE_BUILD_TYPE=RelWithDebInfo \ -DWINDOWS_CERTIFICATE_COMMON_NAME="${WINDOWS_CERTIFICATE_COMMON_NAME}" \ # EOF echo "::endgroup::" echo "::group::Build" cmake --build . echo "::endgroup::" env: WINDOWS_CERTIFICATE_COMMON_NAME: ${{ secrets.WINDOWS_CERTIFICATE_COMMON_NAME }} - name: Create bundles shell: bash run: | cd ${GITHUB_WORKSPACE}/build echo "::group::Run CPack" cpack echo "::endgroup::" echo "::group::Prepare PDB to be bundled" PDB=$(ls bundles/*.zip | cut -d/ -f2 | sed 's/.zip$/.pdb/') cp openttd.pdb bundles/${PDB} xz -9 bundles/${PDB} echo "::endgroup::" echo "::group::Cleanup" # Remove the sha256 files CPack generates; we will do this ourself at # the end of this workflow. rm -f bundles/*.sha256 echo "::endgroup::" - name: Sign installer if: ${{ false }} # needs.source.outputs.is_tag == 'true' shell: bash # If this is run on a fork, there may not be a certificate set up - continue in this case continue-on-error: true run: | cd ${GITHUB_WORKSPACE}/build/bundles ../../os/windows/sign.bat *.exe "${WINDOWS_CERTIFICATE_COMMON_NAME}" env: WINDOWS_CERTIFICATE_COMMON_NAME: ${{ secrets.WINDOWS_CERTIFICATE_COMMON_NAME }} - name: Store bundles uses: actions/upload-artifact@v3 with: name: openttd-windows-${{ matrix.arch }} path: build/bundles retention-days: 5 windows-appx: name: Windows Store needs: - source - windows if: needs.source.outputs.is_tag == 'true' runs-on: windows-latest steps: - name: Download source uses: actions/download-artifact@v3 with: name: internal-source - name: Unpack source shell: bash run: | tar -xf source.tar.gz --strip-components=1 - name: Download x86 build uses: actions/download-artifact@v3 with: name: openttd-windows-x86 - name: Download x64 build uses: actions/download-artifact@v3 with: name: openttd-windows-x64 - name: Download arm64 build uses: actions/download-artifact@v3 with: name: openttd-windows-arm64 - name: Unpack builds shell: bash run: | mkdir builds cd builds function extract { mkdir $1 # Extract the zip version of the release unzip ../openttd-*-windows-$2.zip -d $1 # Remove the extraneous directory mv $1/openttd-*-windows-$2/* $1/ rmdir $1/openttd-*-windows-$2 # Move the openttd.exe to the '{arch}-binaries' folder mkdir $1-binaries mv $1/openttd.exe $1-binaries/ } extract x86 win32 extract x64 win64 extract arm64 arm64 # Use the "x86" folder as the source of the common binaries (lang files, etc) and remove the others mv x86 common-binaries rm -rf x64 arm64 - name: Install OpenGFX shell: bash run: | mkdir -p builds/common-binaries/baseset cd builds/common-binaries/baseset echo "::group::Download OpenGFX" curl -L https://cdn.openttd.org/opengfx-releases/7.1/opengfx-7.1-all.zip -o opengfx-all.zip echo "::endgroup::" echo "::group::Unpack OpenGFX" unzip opengfx-all.zip tar xf opengfx-*.tar echo "::endgroup::" rm -f opengfx-all.zip opengfx-*.tar - name: Install OpenMSX shell: bash run: | mkdir -p builds/common-binaries/baseset cd builds/common-binaries/baseset echo "::group::Download OpenMSX" curl -L https://cdn.openttd.org/openmsx-releases/0.4.2/openmsx-0.4.2-all.zip -o openmsx-all.zip echo "::endgroup::" echo "::group::Unpack OpenGFX" unzip openmsx-all.zip tar xf openmsx-*.tar echo "::endgroup::" rm -f openmsx-all.zip openmsx-*.tar - name: Install OpenSFX shell: bash run: | mkdir -p builds/common-binaries/baseset/opensfx cd builds/common-binaries/baseset/opensfx echo "::group::Download OpenSFX" curl -L https://cdn.openttd.org/opensfx-releases/1.0.3/opensfx-1.0.3-all.zip -o opensfx-all.zip echo "::endgroup::" echo "::group::Unpack OpenSFX" unzip opensfx-all.zip tar xf opensfx-*.tar echo "::endgroup::" rm -f opensfx-all.zip opensfx-*.tar - name: Generate signing certificate shell: cmd run: | cd builds REM We need to provide a signed .appx to the Windows Store, so generate a certificate with password 'password' call ..\os\windows\winstore\generate-key.bat "CN=78024DA8-4BE4-4C77-B12E-547BBF7359D2" password cert.pfx - name: Generate assets shell: cmd run: | cd os\windows\winstore call generate-assets.bat - name: Prepare manifests shell: cmd run: | cd builds mkdir manifests REM Set the version environment variable call ..\os\windows\winstore\set-version.bat x86-binaries\openttd.exe call ..\os\windows\winstore\prepare-manifests.bat manifests "CN=78024DA8-4BE4-4C77-B12E-547BBF7359D2" "57420OpenTTDDevelopers.OpenTTDofficial" - name: Prepare binaries shell: bash run: | cd builds # Copy the Windows Store assets mkdir common-binaries/Assets cp -R ../os/windows/winstore/assets-common/* common-binaries/Assets/ mkdir Assets cp -R ../os/windows/winstore/assets/* Assets/ cp manifests/*.xml . - name: Build shell: cmd run: | REM Add the Windows SDK tools to the PATH echo|set /p="SET VS_INSTALLDIR=" > _vspath.bat vswhere -products * -requires Microsoft.VisualStudio.Component.VC.Tools.x86.x64 -property installationPath >> _vspath.bat call _vspath.bat call "%VS_INSTALLDIR%\Common7\Tools\VsDevCmd.bat" REM Set the version environment variable call os\windows\winstore\set-version.bat builds\x86-binaries\openttd.exe cd builds mkdir output REM Build and sign the package makeappx build /v /f PackagingLayout.xml /op output\ /bv %OTTD_VERSION% /pv %OTTD_VERSION% /ca SignTool sign /fd sha256 /a /f cert.pfx /p password "output\OpenTTD.appxbundle" - name: Store appx uses: actions/upload-artifact@v3 with: name: openttd-windows-store path: | builds/output/OpenTTD.appxbundle builds/cert.pfx retention-days: 5 upload: name: Upload (AWS) needs: - source - docs - linux - linux-distro - macos - windows # The 'linux' job can be skipped if it is a nightly. That normally causes # this job to be skipped too, unless we have this length boy :) # "always()" is important here, it is the keyword to use to stop skipping # this job if any dependency is skipped. It looks a bit silly, but it is # how GitHub Actions work ;) if: ${{ false }} # always() && needs.source.result == 'success' && needs.docs.result == 'success' && needs.linux.result == 'success' && (needs.linux-distro.result == 'success' || needs.linux-distro.result == 'skipped') && needs.macos.result == 'success' && needs.windows.result == 'success' runs-on: ubuntu-20.04 steps: - name: Download all bundles uses: actions/download-artifact@v3 - name: Calculate checksums run: | echo "::group::Move bundles to a single folder" mkdir bundles mv openttd-*/* bundles/ cd bundles echo "::group::Build" for i in $(ls openttd-*); do echo "::group::Calculating checksums for ${i}" openssl dgst -r -md5 -hex $i > $i.md5sum openssl dgst -r -sha1 -hex $i > $i.sha1sum openssl dgst -r -sha256 -hex $i > $i.sha256sum echo "::endgroup::" done - name: Upload bundles to AWS run: | aws s3 cp --recursive --only-show-errors bundles/ s3://${{ secrets.CDN_S3_BUCKET }}/${{ needs.source.outputs.folder }}/${{ needs.source.outputs.version }}/ # We do not invalidate the CloudFront distribution here. The trigger # for "New OpenTTD release" first updated the manifest files and # creates an index.html. We invalidate after that, so everything # becomes visible at once. env: AWS_ACCESS_KEY_ID: ${{ secrets.AWS_ACCESS_KEY_ID }} AWS_SECRET_ACCESS_KEY: ${{ secrets.AWS_SECRET_ACCESS_KEY }} AWS_DEFAULT_REGION: ${{ secrets.AWS_REGION }} - name: Trigger 'New OpenTTD release' uses: peter-evans/repository-dispatch@v2 with: token: ${{ secrets.DEPLOYMENT_TOKEN }} repository: OpenTTD/workflows event-type: ${{ needs.source.outputs.trigger_type }} client-payload: '{"version": "${{ needs.source.outputs.version }}", "folder": "${{ needs.source.outputs.folder }}"}' upload-steam: name: Upload (Steam) needs: - source - linux - macos - windows if: ${{ false }} # needs.source.outputs.trigger_type == 'new-master' || needs.source.outputs.trigger_type == 'new-tag' runs-on: ubuntu-20.04 steps: - name: Download all bundles uses: actions/download-artifact@v3 - name: Setup steamcmd uses: CyberAndrii/setup-steamcmd@v1 - name: Generate Steam auth code id: steam-totp uses: CyberAndrii/steam-totp@v1 with: shared_secret: ${{ secrets.STEAM_SHARED_SECRET }} - name: Upload to Steam run: | echo "::group::Extracting source" mkdir source ( cd source tar -xf ../internal-source/source.tar.gz --strip-components=1 ) echo "::endgroup::" mkdir steam ( cd steam echo "::group::Prepare Win32" unzip ../openttd-windows-x86/openttd-*-windows-win32.zip mv openttd-*-windows-win32 steam-win32 echo "::endgroup::" echo "::group::Prepare Win64" unzip ../openttd-windows-x64/openttd-*-windows-win64.zip mv openttd-*-windows-win64 steam-win64 echo "::endgroup::" echo "::group::Prepare macOS" mkdir steam-macos ( cd steam-macos unzip ../../openttd-macos-universal/openttd-*-macos-universal.zip ) echo "::endgroup::" echo "::group::Prepare Linux" tar xvf ../openttd-linux-generic/openttd-*-linux-generic-amd64.tar.xz mv openttd-*-linux-generic-amd64 steam-linux echo "::endgroup::" echo "::group::Preparing build file" if [ "${{ needs.source.outputs.trigger_type }}" = "new-tag" ]; then BRANCH="testing" else BRANCH="nightly" fi cat ../source/os/steam/release.vdf | sed 's/@@DESCRIPTION@@/openttd-${{ needs.source.outputs.version }}/;s/@@BRANCH@@/'${BRANCH}'/' > release.vdf cat release.vdf echo "::endgroup::" echo "::group::Upload to Steam" steamcmd +login ${{ secrets.STEAM_USERNAME }} ${{ secrets.STEAM_PASSWORD }} ${{ steps.steam-totp.outputs.code }} +run_app_build $(pwd)/release.vdf +quit echo "::endgroup::" )