Skip to content

Commit

Permalink
Merge branch 'feature/reserve_model_lot2_scalian' into feature/reserv…
Browse files Browse the repository at this point in the history
…e_model_lot3_scalian
  • Loading branch information
h-fournier committed Dec 13, 2024
2 parents 2b82698 + 5600547 commit fd2983b
Show file tree
Hide file tree
Showing 11 changed files with 195 additions and 196 deletions.
225 changes: 97 additions & 128 deletions .github/workflows/centos7.yml
Original file line number Diff line number Diff line change
Expand Up @@ -2,11 +2,12 @@ name: Centos7 CI (push and/or release)

on:
release:
types: [created]
types: [ created ]
push:
branches:
- develop
- dependabot/*
- feature/reserve_model_lot3_scalian
schedule:
- cron: '21 2 * * *'
workflow_call:
Expand All @@ -29,135 +30,103 @@ jobs:

build:
name: Build
env:
ORTOOLSDIR: ${{ github.workspace }}/or-tools

runs-on: ubuntu-latest
container: 'antaresrte/rte-antares:centos7-simulator-no-deps'

steps:

- name: Checkout
run: |
git clone $GITHUB_SERVER_URL/$GITHUB_REPOSITORY.git -b ${{ env.REF }} .
- name: Install gcc 10
run: |
# update mirrors, official centos7 is deprecated
sed -i s/mirror.centos.org/vault.centos.org/g /etc/yum.repos.d/*.repo &&\
sed -i s/^#.*baseurl=http/baseurl=http/g /etc/yum.repos.d/*.repo &&\
sed -i s/^mirrorlist=http/#mirrorlist=http/g /etc/yum.repos.d/*.repo &&\
yum update -y
# not a typo, centos-release-scl is needed to install devtoolset-10 but introduce deprecated mirror
sed -i s/mirror.centos.org/vault.centos.org/g /etc/yum.repos.d/*.repo &&\
sed -i s/^#.*baseurl=http/baseurl=http/g /etc/yum.repos.d/*.repo &&\
sed -i s/^mirrorlist=http/#mirrorlist=http/g /etc/yum.repos.d/*.repo
yum install -y centos-release-scl
yum install -y devtoolset-11-gcc*
- name: Install cmake 3.28
run: pip3 install cmake==3.28.4

- name: Install VCPKG
# Note: we need to use environment variables instead of workflow variables
# because github messes up path variables when running in container,
# see https://github.com/actions/runner/issues/2058
run: |
git submodule update --init vcpkg && ./vcpkg/bootstrap-vcpkg.sh -disableMetrics
echo "VCPKG_ROOT=$GITHUB_WORKSPACE/vcpkg" >> $GITHUB_ENV
echo "VCPKG_CACHE_DIR=$GITHUB_WORKSPACE/vcpkg_cache" >> $GITHUB_ENV
echo "VCPKG_BINARY_SOURCES=clear;files,$GITHUB_WORKSPACE/vcpkg_cache,readwrite" >> $GITHUB_ENV
- name: Restore vcpkg binary dir from cache
id: cache-vcpkg-binary
# Note: we are stuck with v3, because v4 is not compatible with oracle8 image
uses: actions/cache/restore@v3
with:
path: ${{ env.VCPKG_CACHE_DIR }}
key: vcpkg-cache-centos7-${{ hashFiles('src/vcpkg.json', '.git/modules/vcpkg/HEAD') }}
# Allows to restore a cache when deps have only partially changed (like adding a dependency)
restore-keys: vcpkg-cache-centos7-

- name: Init submodule
run: |
git submodule update --init --remote src/tests/resources/Antares_Simulator_Tests
- name: Config OR-Tools URL
run: |
echo "URL_ORTOOLS=https://github.com/rte-france/or-tools-rte/releases/download/$(cat ortools_tag)/ortools_cxx_centos7_static_sirius.zip" >> $GITHUB_ENV
- name: Download OR-Tools
id: ortools
run: |
mkdir -p ${{ env.ORTOOLSDIR }} && cd ${{ env.ORTOOLSDIR }}
wget -q -O ortools.zip ${{ env.URL_ORTOOLS }}
unzip -q ortools.zip
rm ortools.zip
- name: Install gh if needed
if: ${{ env.IS_RELEASE == 'true' }}
run: |
wget https://github.com/cli/cli/releases/download/v2.52.0/gh_2.52.0_linux_amd64.rpm
rpm -i gh_2.52.0_linux_amd64.rpm
gh --version
- name: Configure
run: |
source /opt/rh/devtoolset-11/enable
source /opt/rh/rh-git227/enable
cmake -B _build -S src \
-DCMAKE_C_COMPILER_LAUNCHER=ccache \
-DCMAKE_CXX_COMPILER_LAUNCHER=ccache \
-DCMAKE_TOOLCHAIN_FILE=$GITHUB_WORKSPACE/vcpkg/scripts/buildsystems/vcpkg.cmake \
-DVCPKG_TARGET_TRIPLET=x64-linux-release \
-DCMAKE_BUILD_TYPE=Release \
-DBUILD_TESTING=ON \
-DBUILD_TOOLS=ON \
-DBUILD_UI=OFF \
-DCMAKE_PREFIX_PATH=${{ env.ORTOOLSDIR }}/install \
- name: Build
run: |
source /opt/rh/devtoolset-11/enable
source /opt/rh/rh-git227/enable
cmake --build _build --config Release -j$(nproc)
ccache -s
- name: Installer .rpm creation
run: |
cd _build
cpack -G RPM
- name: Solver archive creation
run: |
cd _build
cmake --install . --prefix install
pushd .
cd install/bin
tar czf ../../antares-solver_centos7.tar.gz antares-solver libsirius_solver.so
popd
rm -rf install
- name: .tar.gz creation
run: |
cd _build
cpack -G TGZ
- name: Publish assets
if: ${{ env.IS_RELEASE == 'true' }}
env:
GITHUB_TOKEN: ${{ github.token }}
tag: ${{ github.event.inputs.release_tag }}
run: |
gh release upload "$tag" _build/*.tar.gz _build/*.rpm
- name: Cache vcpkg binary dir
if: always()
id: save-cache-vcpkg-binary
uses: actions/cache/save@v3
with:
path: ${{ env.VCPKG_CACHE_DIR }}
key: vcpkg-cache-centos7-${{ hashFiles('src/vcpkg.json', '.git/modules/vcpkg/HEAD') }}

#sparse checkout -- only needed files
- uses: actions/checkout@v4
with:
fetch-depth: 0
sparse-checkout: |
docker
src/vcpkg.json
ref: ${{ env.REF }}

- name: set env variables -- DockerFiles
run: |
echo "DOCKERFILE=$(pwd)/docker/Dockerfile" >> $GITHUB_ENV
echo "DOCKERDIR=$(pwd)/docker" >> $GITHUB_ENV
- name: Get changed files
id: changed-files
uses: tj-actions/changed-files@v45
with:
files: |
docker/AntaresDeps
- name: Docker file push
id: docker_push

if: steps.changed-files.outputs.any_changed == 'true'
uses: elgohr/Publish-Docker-Github-Action@main
with:
name: antaresrte/antaressystemdeps
username: ${{ secrets.DOCKER_USERNAME }}
password: ${{ secrets.DOCKER_PASSWORD }}
workdir: ${{ env.DOCKERDIR }}
dockerfile: ${{ env.DOCKERDIR }}/AntaresDeps
cache: false
tags: latest

- name: create vcpkg cache dir
run: |
echo "VCPKG_CACHE_DIR=$GITHUB_WORKSPACE/vcpkg_cache" >> $GITHUB_ENV
mkdir -p ${{ github.workspace }}/vcpkg_cache
- name: Restore vcpkg binary dir from cache
id: cache-vcpkg-binary
uses: actions/cache/restore@v4
with:
path: ${{ github.workspace }}/vcpkg_cache
key: vcpkg-cache-centos7-${{ hashFiles('src/vcpkg.json', '.git/modules/vcpkg/HEAD') }}
# Allows to restore a cache when deps have only partially changed (like adding a dependency)
restore-keys: vcpkg-cache-centos7-

- name: Setup ccache
uses: hendrikmuhs/[email protected]
with:
key: centos7-on-${{ runner.os }}


- name: Build the image and Antares
run: |
ls -latr .ccache/ && \
docker build \
-t antares:centos7 \
--build-arg="BRANCH=${{ env.REF }}" \
--build-arg="NPROC=$(nproc)" \
--build-arg="VCPKG_CACHE_DIR=./vcpkg_cache" \
--build-arg CCACHE_DIR=./.ccache \
--build-arg CCACHE_KEY=centos7-on-${{ runner.os }} \
-f ${{ env.DOCKERFILE }} .
- name: create a container without starting it && retrieve the .tgz
run: |
container_id=$(docker create antares:centos7)
docker cp $container_id:/workspace/Antares_Simulator/_build/archive archive
docker cp $container_id:/workspace/vcpkg_cache ${{ env.VCPKG_CACHE_DIR }}
docker cp $container_id:/workspace/.ccache/. .ccache
ls -la .ccache
docker rm $container_id
- name: Publish assets
if: ${{ env.IS_RELEASE == 'true' }}
env:
GITHUB_TOKEN: ${{ github.token }}
tag: ${{ github.event.inputs.release_tag }}
run: |
gh release upload "$tag" archive/*.tar.gz archive/*.rpm

- name: Upload .tar.gz
uses: actions/upload-artifact@v4
with:
path: archive/*.tar.gz
name: antares-9.2.0-rc4-CentOS-7.9.2009.tar.gz


- name: Cache vcpkg binary dir
if: always()
id: save-cache-vcpkg-binary
uses: actions/cache/save@v4
with:
path: ${{ github.workspace }}/vcpkg_cache
key: vcpkg-cache-centos7-${{ hashFiles('src/vcpkg.json', '.git/modules/vcpkg/HEAD') }}
1 change: 1 addition & 0 deletions src/antares-deps
Submodule antares-deps added at 0d6beb
2 changes: 1 addition & 1 deletion src/libs/antares/study/include/antares/study/area/area.h
Original file line number Diff line number Diff line change
Expand Up @@ -87,7 +87,7 @@ class BiMap
}

// Function to return the size of the bimap
int size()
int size() const
{
return key_to_value.size();
}
Expand Down
30 changes: 24 additions & 6 deletions src/libs/antares/study/parts/common/cluster_list.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -437,12 +437,30 @@ bool ClusterList<ClusterT>::loadReserveParticipations(Area& area, const std::fil
auto cluster = area.thermal.list.getClusterByName(tmpClusterName);
if (reserve && cluster)
{
ThermalClusterReserveParticipation tmpReserveParticipation{reserve.value(),
tmpMaxPower,
tmpParticipationCost,
tmpMaxPowerOff,
tmpParticipationCostOff};
cluster.value().get()->addReserveParticipation(section.name, tmpReserveParticipation);
bool isClusterMustRun = false;
for (const auto& clusterMustRun : area.thermal.list.each_mustrun_and_enabled())
{
if (clusterMustRun->id() == cluster.value().get()->id())
{
isClusterMustRun = true;
break;
}
}
if (!isClusterMustRun)
{
ThermalClusterReserveParticipation tmpReserveParticipation{ reserve.value(),
tmpMaxPower,
tmpParticipationCost,
tmpMaxPowerOff,
tmpParticipationCostOff };


cluster.value().get()->addReserveParticipation(section.name, tmpReserveParticipation);
}
else
{
logs.warning() << area.name << " : " << tmpClusterName << " is mustrun and is participating in capacity reservation " << section.name << ", ignored.";
}
}
else
{
Expand Down
15 changes: 6 additions & 9 deletions src/solver/optimisation/opt_gestion_second_membre_reserves.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -46,15 +46,12 @@ void OPT_InitialiserLeSecondMembreDuProblemeLineaireReserves(PROBLEME_HEBDO* pro
for (int pdtJour = 0, pdtHebdo = PremierPdtDeLIntervalle; pdtHebdo < DernierPdtDeLIntervalle;
pdtHebdo++, pdtJour++)
{
int hourInTheYear = problemeHebdo->weekInTheYear * 168 + pdtHebdo;
const CORRESPONDANCES_DES_CONTRAINTES& CorrespondanceCntNativesCntOptim
= problemeHebdo->CorrespondanceCntNativesCntOptim[pdtJour];

for (uint32_t pays = 0; pays < problemeHebdo->NombreDePays; pays++)
{
int pdtGlobal = problemeHebdo->weekInTheYear
* problemeHebdo->NombreDePasDeTempsDUneJournee
* problemeHebdo->NombreDeJours
+ pdtJour;

// Thermal clusters
{
Expand All @@ -67,7 +64,7 @@ void OPT_InitialiserLeSecondMembreDuProblemeLineaireReserves(PROBLEME_HEBDO* pro
[areaReserveUp.globalReserveIndex];
if (cnt >= 0)
{
SecondMembre[cnt] = areaReserveUp.need.at(pdtGlobal);
SecondMembre[cnt] = areaReserveUp.need.at(hourInTheYear);
double* adresseDuResultat = &(problemeHebdo->ResultatsHoraires[pays].Reserves[pdtHebdo].CoutsMarginauxHoraires[areaReserveUp.areaReserveIndex]);
AdresseOuPlacerLaValeurDesCoutsMarginaux[cnt] = adresseDuResultat;
}
Expand Down Expand Up @@ -100,7 +97,7 @@ void OPT_InitialiserLeSecondMembreDuProblemeLineaireReserves(PROBLEME_HEBDO* pro
[areaReserveDown.globalReserveIndex];
if (cnt >= 0)
{
SecondMembre[cnt] = areaReserveDown.need.at(pdtGlobal);
SecondMembre[cnt] = areaReserveDown.need.at(hourInTheYear);
double* adresseDuResultat = &(problemeHebdo->ResultatsHoraires[pays].Reserves[pdtHebdo].CoutsMarginauxHoraires[areaReserveDown.areaReserveIndex]);
AdresseOuPlacerLaValeurDesCoutsMarginaux[cnt] = adresseDuResultat;
}
Expand Down Expand Up @@ -234,7 +231,7 @@ void OPT_InitialiserLeSecondMembreDuProblemeLineaireReserves(PROBLEME_HEBDO* pro
[globalClusterIdx];
if (cnt >= 0)
{
SecondMembre[cnt] = cluster.series.get()->maxWithdrawalModulation[pdtJour]
SecondMembre[cnt] = cluster.series.get()->maxWithdrawalModulation[hourInTheYear]
* cluster.withdrawalNominalCapacity;
AdresseOuPlacerLaValeurDesCoutsMarginaux[cnt] = nullptr;
}
Expand All @@ -245,7 +242,7 @@ void OPT_InitialiserLeSecondMembreDuProblemeLineaireReserves(PROBLEME_HEBDO* pro
[globalClusterIdx];
if (cnt >= 0)
{
SecondMembre[cnt] = cluster.series.get()->lowerRuleCurve[pdtJour]
SecondMembre[cnt] = cluster.series.get()->lowerRuleCurve[hourInTheYear]
* cluster.withdrawalNominalCapacity;
AdresseOuPlacerLaValeurDesCoutsMarginaux[cnt] = nullptr;
}
Expand All @@ -256,7 +253,7 @@ void OPT_InitialiserLeSecondMembreDuProblemeLineaireReserves(PROBLEME_HEBDO* pro
[globalClusterIdx];
if (cnt >= 0)
{
SecondMembre[cnt] = cluster.series.get()->maxInjectionModulation[pdtJour]
SecondMembre[cnt] = cluster.series.get()->maxInjectionModulation[hourInTheYear]
* cluster.injectionNominalCapacity;
AdresseOuPlacerLaValeurDesCoutsMarginaux[cnt] = nullptr;
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -106,7 +106,7 @@ class ReserveParticipationByDispatchableOffUnitsPlant
pNbYearsParallel = study->maxNbYearsInParallel;
pValuesForTheCurrentYear = new VCardType::IntermediateValuesBaseType[pNbYearsParallel];

// Get the area
// Get the number of Onthermal reserveParticipations
pSize = area->thermal.list.reserveParticipationsCount();

if (pSize)
Expand Down Expand Up @@ -266,12 +266,20 @@ class ReserveParticipationByDispatchableOffUnitsPlant
// Write the data for the current year
for (uint i = 0; i < pSize; ++i)
{
auto [clusterName, reserveName]
= results.data.area->reserveParticipationThermalClustersIndexMap.get(i);
results.variableCaption = clusterName + "_" + reserveName; // VCardType::Caption();
results.variableUnit = VCardType::Unit();
pValuesForTheCurrentYear[numSpace][i].template buildAnnualSurveyReport<VCardType>(
results, fileLevel, precision);
if (results.data.area->reserveParticipationThermalClustersIndexMap.size() == 0) //Bimap is empty
{
logs.warning() << "Problem during the results export, the thermal bimap is empty for area " << results.data.area->name;
break;
}
else
{
auto [clusterName, reserveName]
= results.data.area->reserveParticipationThermalClustersIndexMap.get(i);
results.variableCaption = clusterName + "_" + reserveName; // VCardType::Caption();
results.variableUnit = VCardType::Unit();
pValuesForTheCurrentYear[numSpace][i].template buildAnnualSurveyReport<VCardType>(
results, fileLevel, precision);
}
}
}
}
Expand Down
Loading

0 comments on commit fd2983b

Please sign in to comment.