Files
ortools-clone/ortools/sat/scheduling_helpers.cc
Mizux Seiha 4f381f6d07 backport from main:
* bump abseil to 20250814
* bump protobuf to v32.0
* cmake: add ccache auto support
* backport flatzinc, math_opt and sat update
2025-09-16 16:25:04 +02:00

1200 lines
44 KiB
C++

// Copyright 2010-2025 Google LLC
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#include "ortools/sat/scheduling_helpers.h"
#include <algorithm>
#include <functional>
#include <string>
#include <utility>
#include <vector>
#include "absl/log/check.h"
#include "absl/strings/str_cat.h"
#include "absl/types/span.h"
#include "ortools/base/logging.h"
#include "ortools/sat/enforcement.h"
#include "ortools/sat/implied_bounds.h"
#include "ortools/sat/integer.h"
#include "ortools/sat/integer_base.h"
#include "ortools/sat/integer_expr.h"
#include "ortools/sat/linear_constraint.h"
#include "ortools/sat/model.h"
#include "ortools/sat/precedences.h"
#include "ortools/sat/sat_base.h"
#include "ortools/sat/sat_solver.h"
#include "ortools/util/sort.h"
#include "ortools/util/strong_integers.h"
namespace operations_research {
namespace sat {
SchedulingConstraintHelper::SchedulingConstraintHelper(
std::vector<AffineExpression> starts, std::vector<AffineExpression> ends,
std::vector<AffineExpression> sizes,
std::vector<LiteralIndex> reason_for_presence, Model* model)
: model_(model),
sat_solver_(model->GetOrCreate<SatSolver>()),
assignment_(sat_solver_->Assignment()),
trail_(model->GetOrCreate<Trail>()),
integer_trail_(model->GetOrCreate<IntegerTrail>()),
watcher_(model->GetOrCreate<GenericLiteralWatcher>()),
linear2_bounds_(model->GetOrCreate<Linear2Bounds>()),
root_level_lin2_bounds_(model->GetOrCreate<RootLevelLinear2Bounds>()),
enforcement_helper_(*model->GetOrCreate<EnforcementHelper>()),
enforcement_id_(-1),
starts_(std::move(starts)),
ends_(std::move(ends)),
sizes_(std::move(sizes)),
reason_for_presence_(std::move(reason_for_presence)),
capacity_(starts_.size()),
cached_size_min_(new IntegerValue[capacity_]),
cached_start_min_(new IntegerValue[capacity_]),
cached_end_min_(new IntegerValue[capacity_]),
cached_negated_start_max_(new IntegerValue[capacity_]),
cached_negated_end_max_(new IntegerValue[capacity_]),
cached_shifted_start_min_(new IntegerValue[capacity_]),
cached_negated_shifted_end_max_(new IntegerValue[capacity_]) {
DCHECK_EQ(starts_.size(), ends_.size());
DCHECK_EQ(starts_.size(), sizes_.size());
DCHECK_EQ(starts_.size(), reason_for_presence_.size());
minus_starts_.clear();
minus_starts_.reserve(starts_.size());
minus_ends_.clear();
minus_ends_.reserve(starts_.size());
for (int i = 0; i < starts_.size(); ++i) {
minus_starts_.push_back(starts_[i].Negated());
minus_ends_.push_back(ends_[i].Negated());
}
InitSortedVectors();
if (!SynchronizeAndSetTimeDirection(true)) {
model->GetOrCreate<SatSolver>()->NotifyThatModelIsUnsat();
}
}
SchedulingConstraintHelper::SchedulingConstraintHelper(int num_tasks,
Model* model)
: model_(model),
sat_solver_(model->GetOrCreate<SatSolver>()),
assignment_(sat_solver_->Assignment()),
trail_(model->GetOrCreate<Trail>()),
integer_trail_(model->GetOrCreate<IntegerTrail>()),
linear2_bounds_(model->GetOrCreate<Linear2Bounds>()),
root_level_lin2_bounds_(model->GetOrCreate<RootLevelLinear2Bounds>()),
enforcement_helper_(*model->GetOrCreate<EnforcementHelper>()),
enforcement_id_(-1),
capacity_(num_tasks),
cached_size_min_(new IntegerValue[capacity_]),
cached_start_min_(new IntegerValue[capacity_]),
cached_end_min_(new IntegerValue[capacity_]),
cached_negated_start_max_(new IntegerValue[capacity_]),
cached_negated_end_max_(new IntegerValue[capacity_]),
cached_shifted_start_min_(new IntegerValue[capacity_]),
cached_negated_shifted_end_max_(new IntegerValue[capacity_]) {
starts_.resize(num_tasks);
CHECK_EQ(NumTasks(), num_tasks);
}
bool SchedulingConstraintHelper::IsEnforced() const {
return enforcement_helper_.Status(enforcement_id_) ==
EnforcementStatus::IS_ENFORCED;
}
bool SchedulingConstraintHelper::Propagate() {
if (!IsEnforced()) return true;
recompute_all_cache_ = true;
for (const int id : propagator_ids_) watcher_->CallOnNextPropagate(id);
return true;
}
bool SchedulingConstraintHelper::IncrementalPropagate(
const std::vector<int>& watch_indices) {
if (!IsEnforced()) return true;
for (const int t : watch_indices) recompute_cache_.Set(t);
for (const int id : propagator_ids_) watcher_->CallOnNextPropagate(id);
return true;
}
void SchedulingConstraintHelper::RegisterWith(
GenericLiteralWatcher* watcher,
absl::Span<const Literal> enforcement_literals) {
const int id = watcher->Register(this);
const int num_tasks = starts_.size();
for (int t = 0; t < num_tasks; ++t) {
watcher->WatchIntegerVariable(sizes_[t].var, id, t);
watcher->WatchIntegerVariable(starts_[t].var, id, t);
watcher->WatchIntegerVariable(ends_[t].var, id, t);
// This class do not need to be waked up on presence change, since this is
// not cached. However given that we can have many propagators that use the
// same helper, it is nicer to only register this one, and wake up all
// propagator through it rather than registering all of them individually.
// Note that IncrementalPropagate() will do nothing if this is the only
// change except waking up registered propagators.
if (!IsPresent(t) && !IsAbsent(t)) {
watcher->WatchLiteral(Literal(reason_for_presence_[t]), id);
}
}
watcher->SetPropagatorPriority(id, 0);
enforcement_id_ =
enforcement_helper_.Register(enforcement_literals, watcher, id);
}
bool SchedulingConstraintHelper::UpdateCachedValues(int t) {
if (IsAbsent(t)) return true;
IntegerValue smin = integer_trail_->LowerBound(starts_[t]);
IntegerValue smax = integer_trail_->UpperBound(starts_[t]);
IntegerValue emin = integer_trail_->LowerBound(ends_[t]);
IntegerValue emax = integer_trail_->UpperBound(ends_[t]);
// We take the max for the corner case where the size of an optional interval
// is used elsewhere and has a domain with negative value.
//
// TODO(user): maybe we should just disallow size with a negative domain, but
// is is harder to enforce if we have a linear expression for size.
IntegerValue dmin =
std::max(IntegerValue(0), integer_trail_->LowerBound(sizes_[t]));
IntegerValue dmax = integer_trail_->UpperBound(sizes_[t]);
// Detect first if we have a conflict using the relation start + size = end.
if (dmax < 0) {
ResetReason();
AddSizeMaxReason(t, dmax);
return PushTaskAbsence(t);
}
if (smin + dmin - emax > 0) {
ResetReason();
AddStartMinReason(t, smin);
AddSizeMinReason(t, dmin);
AddEndMaxReason(t, emax);
return PushTaskAbsence(t);
}
if (smax + dmax - emin < 0) {
ResetReason();
AddStartMaxReason(t, smax);
AddSizeMaxReason(t, dmax);
AddEndMinReason(t, emin);
return PushTaskAbsence(t);
}
// Sometimes, for optional interval with non-optional bounds, this propagation
// give tighter bounds. We always consider the value assuming
// the interval is present.
//
// Note that this is also useful in case not everything was propagated. Note
// also that since there is no conflict, we reach the fix point in one pass.
smin = std::max(smin, emin - dmax);
smax = std::min(smax, emax - dmin);
dmin = std::max(dmin, emin - smax);
emin = std::max(emin, smin + dmin);
emax = std::min(emax, smax + dmax);
if (emin != cached_end_min_[t]) {
recompute_energy_profile_ = true;
}
// We might only want to do that if the value changed, but I am not sure it
// is worth the test.
recompute_by_start_max_ = true;
recompute_by_end_min_ = true;
cached_start_min_[t] = smin;
cached_end_min_[t] = emin;
cached_negated_start_max_[t] = -smax;
cached_negated_end_max_[t] = -emax;
cached_size_min_[t] = dmin;
// Note that we use the cached value here for EndMin()/StartMax().
const IntegerValue new_shifted_start_min = emin - dmin;
if (new_shifted_start_min != cached_shifted_start_min_[t]) {
recompute_energy_profile_ = true;
recompute_shifted_start_min_ = true;
cached_shifted_start_min_[t] = new_shifted_start_min;
}
const IntegerValue new_negated_shifted_end_max = -(smax + dmin);
if (new_negated_shifted_end_max != cached_negated_shifted_end_max_[t]) {
recompute_negated_shifted_end_max_ = true;
cached_negated_shifted_end_max_[t] = new_negated_shifted_end_max;
}
return true;
}
bool SchedulingConstraintHelper::ResetFromSubset(
const SchedulingConstraintHelper& other, absl::Span<const int> tasks) {
current_time_direction_ = other.current_time_direction_;
const int num_tasks = tasks.size();
CHECK_LE(num_tasks, capacity_);
starts_.resize(num_tasks);
ends_.resize(num_tasks);
minus_ends_.resize(num_tasks);
minus_starts_.resize(num_tasks);
sizes_.resize(num_tasks);
reason_for_presence_.resize(num_tasks);
for (int i = 0; i < num_tasks; ++i) {
const int t = tasks[i];
starts_[i] = other.starts_[t];
ends_[i] = other.ends_[t];
minus_ends_[i] = other.minus_ends_[t];
minus_starts_[i] = other.minus_starts_[t];
sizes_[i] = other.sizes_[t];
reason_for_presence_[i] = other.reason_for_presence_[t];
}
InitSortedVectors();
return SynchronizeAndSetTimeDirection(true);
}
void SchedulingConstraintHelper::InitSortedVectors() {
const int num_tasks = starts_.size();
recompute_all_cache_ = true;
recompute_cache_.Resize(num_tasks);
non_fixed_intervals_.resize(num_tasks);
for (int t = 0; t < num_tasks; ++t) {
non_fixed_intervals_[t] = t;
recompute_cache_.Set(t);
}
// Make sure all the cached_* arrays can hold enough data.
CHECK_LE(num_tasks, capacity_);
task_by_increasing_start_min_.resize(num_tasks);
task_by_increasing_end_min_.resize(num_tasks);
task_by_increasing_negated_start_max_.resize(num_tasks);
task_by_decreasing_end_max_.resize(num_tasks);
task_by_increasing_shifted_start_min_.resize(num_tasks);
task_by_negated_shifted_end_max_.resize(num_tasks);
for (int t = 0; t < num_tasks; ++t) {
task_by_increasing_start_min_[t].task_index = t;
task_by_increasing_end_min_[t].task_index = t;
task_by_increasing_negated_start_max_[t].task_index = t;
task_by_decreasing_end_max_[t].task_index = t;
task_by_increasing_shifted_start_min_[t].task_index = t;
task_by_increasing_shifted_start_min_[t].presence_lit =
reason_for_presence_[t];
task_by_negated_shifted_end_max_[t].task_index = t;
task_by_negated_shifted_end_max_[t].presence_lit = reason_for_presence_[t];
}
recompute_by_start_max_ = true;
recompute_by_end_min_ = true;
recompute_energy_profile_ = true;
recompute_shifted_start_min_ = true;
recompute_negated_shifted_end_max_ = true;
}
void SchedulingConstraintHelper::SetTimeDirection(bool is_forward) {
if (current_time_direction_ != is_forward) {
current_time_direction_ = is_forward;
std::swap(starts_, minus_ends_);
std::swap(ends_, minus_starts_);
std::swap(task_by_increasing_start_min_, task_by_decreasing_end_max_);
std::swap(task_by_increasing_end_min_,
task_by_increasing_negated_start_max_);
std::swap(recompute_by_end_min_, recompute_by_start_max_);
std::swap(task_by_increasing_shifted_start_min_,
task_by_negated_shifted_end_max_);
recompute_energy_profile_ = true;
std::swap(cached_start_min_, cached_negated_end_max_);
std::swap(cached_end_min_, cached_negated_start_max_);
std::swap(cached_shifted_start_min_, cached_negated_shifted_end_max_);
std::swap(recompute_shifted_start_min_, recompute_negated_shifted_end_max_);
}
}
bool SchedulingConstraintHelper::SynchronizeAndSetTimeDirection(
bool is_forward) {
SetTimeDirection(is_forward);
// If there was any backtracks since the last time this was called, we
// recompute our cache.
if (sat_solver_->num_backtracks() != saved_num_backtracks_) {
recompute_all_cache_ = true;
saved_num_backtracks_ = sat_solver_->num_backtracks();
}
if (recompute_all_cache_) {
for (const int t : non_fixed_intervals_) {
if (!UpdateCachedValues(t)) return false;
}
// We also update non_fixed_intervals_ at level zero so that we will never
// scan them again.
if (sat_solver_->CurrentDecisionLevel() == 0) {
int new_size = 0;
for (const int t : non_fixed_intervals_) {
if (IsPresent(t) && StartIsFixed(t) && EndIsFixed(t) &&
SizeIsFixed(t)) {
continue;
}
non_fixed_intervals_[new_size++] = t;
}
non_fixed_intervals_.resize(new_size);
}
} else {
for (const int t : recompute_cache_) {
if (!UpdateCachedValues(t)) return false;
}
}
recompute_cache_.ClearAll();
recompute_all_cache_ = false;
return true;
}
IntegerValue SchedulingConstraintHelper::GetCurrentMinDistanceBetweenTasks(
int a, int b) {
const AffineExpression before = ends_[a];
const AffineExpression after = starts_[b];
const LinearExpression2 expr(before.var, after.var, before.coeff,
-after.coeff);
const IntegerValue expr_ub = linear2_bounds_->UpperBound(expr);
const IntegerValue needed_offset = before.constant - after.constant;
const IntegerValue ub_of_end_minus_start = expr_ub + needed_offset;
const IntegerValue distance = -ub_of_end_minus_start;
return distance;
}
// Note that we could call this at a positive level to propagate any literal
// associated to task a before task b. However we only call this for task that
// are in detectable precedence, which means the normal precedence or linear
// propagator should have already propagated that Boolean too.
bool SchedulingConstraintHelper::NotifyLevelZeroPrecedence(int a, int b) {
CHECK(IsPresent(a));
CHECK(IsPresent(b));
CHECK_EQ(sat_solver_->CurrentDecisionLevel(), 0);
// Convert ends_[a] <= starts[b] to linear2 <= rhs and canonicalize.
const auto [expr, rhs] = EncodeDifferenceLowerThan(ends_[a], starts_[b], 0);
// Trivial case.
if (expr.coeffs[0] == 0 && expr.coeffs[1] == 0) {
if (rhs < 0) {
sat_solver_->NotifyThatModelIsUnsat();
return false;
}
return true;
}
if (root_level_lin2_bounds_->AddUpperBound(expr, rhs)) {
VLOG(2) << "new relation " << TaskDebugString(a)
<< " <= " << TaskDebugString(b);
// TODO(user): Adding new constraint during propagation might not be the
// best idea as it can create some complication.
AddWeightedSumLowerOrEqual({}, {expr.vars[0], expr.vars[1]},
{expr.coeffs[0].value(), expr.coeffs[1].value()},
rhs.value(), model_);
if (sat_solver_->ModelIsUnsat()) return false;
}
return true;
}
absl::Span<const TaskTime>
SchedulingConstraintHelper::TaskByIncreasingStartMin() {
for (TaskTime& ref : task_by_increasing_start_min_) {
ref.time = StartMin(ref.task_index);
}
IncrementalSort(task_by_increasing_start_min_.begin(),
task_by_increasing_start_min_.end());
return task_by_increasing_start_min_;
}
absl::Span<const TaskTime>
SchedulingConstraintHelper::TaskByIncreasingEndMin() {
if (!recompute_by_end_min_) return task_by_increasing_end_min_;
for (TaskTime& ref : task_by_increasing_end_min_) {
ref.time = EndMin(ref.task_index);
}
IncrementalSort(task_by_increasing_end_min_.begin(),
task_by_increasing_end_min_.end());
recompute_by_end_min_ = false;
return task_by_increasing_end_min_;
}
absl::Span<const TaskTime>
SchedulingConstraintHelper::TaskByIncreasingNegatedStartMax() {
if (!recompute_by_start_max_) return task_by_increasing_negated_start_max_;
for (TaskTime& ref : task_by_increasing_negated_start_max_) {
ref.time = cached_negated_start_max_[ref.task_index];
}
IncrementalSort(task_by_increasing_negated_start_max_.begin(),
task_by_increasing_negated_start_max_.end());
recompute_by_start_max_ = false;
return task_by_increasing_negated_start_max_;
}
absl::Span<const TaskTime>
SchedulingConstraintHelper::TaskByDecreasingEndMax() {
for (TaskTime& ref : task_by_decreasing_end_max_) {
ref.time = EndMax(ref.task_index);
}
IncrementalSort(task_by_decreasing_end_max_.begin(),
task_by_decreasing_end_max_.end(), std::greater<TaskTime>());
return task_by_decreasing_end_max_;
}
absl::Span<const CachedTaskBounds>
SchedulingConstraintHelper::TaskByIncreasingShiftedStartMin() {
if (recompute_shifted_start_min_) {
recompute_shifted_start_min_ = false;
bool is_sorted = true;
IntegerValue previous = kMinIntegerValue;
for (CachedTaskBounds& ref : task_by_increasing_shifted_start_min_) {
ref.time = ShiftedStartMin(ref.task_index);
is_sorted = is_sorted && ref.time >= previous;
previous = ref.time;
}
if (is_sorted) return task_by_increasing_shifted_start_min_;
IncrementalSort(task_by_increasing_shifted_start_min_.begin(),
task_by_increasing_shifted_start_min_.end());
}
return task_by_increasing_shifted_start_min_;
}
absl::Span<const CachedTaskBounds>
SchedulingConstraintHelper::TaskByIncreasingNegatedShiftedEndMax() {
if (recompute_negated_shifted_end_max_) {
recompute_negated_shifted_end_max_ = false;
bool is_sorted = true;
IntegerValue previous = kMinIntegerValue;
for (CachedTaskBounds& ref : task_by_negated_shifted_end_max_) {
ref.time = -ShiftedEndMax(ref.task_index);
is_sorted = is_sorted && ref.time >= previous;
previous = ref.time;
}
if (is_sorted) return task_by_negated_shifted_end_max_;
IncrementalSort(task_by_negated_shifted_end_max_.begin(),
task_by_negated_shifted_end_max_.end());
}
return task_by_negated_shifted_end_max_;
}
// TODO(user): Avoid recomputing it if nothing changed.
const std::vector<SchedulingConstraintHelper::ProfileEvent>&
SchedulingConstraintHelper::GetEnergyProfile() {
if (energy_profile_.empty()) {
const int num_tasks = NumTasks();
for (int t = 0; t < num_tasks; ++t) {
energy_profile_.push_back(
{cached_shifted_start_min_[t], t, /*is_first=*/true});
energy_profile_.push_back({cached_end_min_[t], t, /*is_first=*/false});
}
} else {
if (!recompute_energy_profile_) return energy_profile_;
for (ProfileEvent& ref : energy_profile_) {
const int t = ref.task;
if (ref.is_first) {
ref.time = cached_shifted_start_min_[t];
} else {
ref.time = cached_end_min_[t];
}
}
}
IncrementalSort(energy_profile_.begin(), energy_profile_.end());
recompute_energy_profile_ = false;
return energy_profile_;
}
bool SchedulingConstraintHelper::TaskIsBeforeOrIsOverlapping(int before,
int after) {
const auto [expr, ub] =
EncodeDifferenceLowerThan(starts_[before], ends_[after], -1);
return linear2_bounds_->UpperBound(expr) <= ub;
}
void SchedulingConstraintHelper::AddReasonForBeingBeforeAssumingNoOverlap(
int before, int after) {
FlagItemAsUsedInReason(before);
FlagItemAsUsedInReason(after);
// We compute this as an optimization, since for fixed sizes all linear2
// options are equivalent.
const bool fixed_size = sizes_[before].var == kNoIntegerVariable &&
sizes_[after].var == kNoIntegerVariable &&
starts_[before].var == ends_[before].var &&
starts_[after].var == ends_[after].var;
// Prefer the straightforward linear2 explanation as it is more likely this
// comes from level zero or a single enforcement literal. Also handle the
// fixed size case. This explains with at most two integer bounds.
{
const auto [expr, ub] =
EncodeDifferenceLowerThan(starts_[before], ends_[after], -1);
if (fixed_size || linear2_bounds_->UpperBound(expr) <= ub) {
linear2_bounds_->AddReasonForUpperBoundLowerThan(
expr, ub, &literal_reason_, &integer_reason_);
return;
}
}
// Another choice of Linear2. We need Start(before) < End(after). We rewrite
// as:
// End(before) - Size(before) < Start(after) + Size(after)
//
// Note that the generic code below not based on linear2 will not work
// if we can only get a good bound for End(before)-Start(after), so we also
// handle it here even if the linear2 is not known.
//
// TODO(user): check also the linear2 constructed with (end_before,
// end_after) and (start_after, start_before). Or maybe keep a index of pair
// of intervals that have non-trivial linear2 bounds and use that instead.
{
const auto [expr, ub] = EncodeDifferenceLowerThan(
ends_[before], starts_[after], SizeMin(before) + SizeMin(after) - 1);
if (linear2_bounds_->UpperBound(expr) <= ub) {
AddSizeMinReason(before);
AddSizeMinReason(after);
linear2_bounds_->AddReasonForUpperBoundLowerThan(
expr, ub, &literal_reason_, &integer_reason_);
return;
}
}
// We will explain StartMax(before) < EndMin(after);
DCHECK_LT(StartMax(before), EndMin(after));
// The reason will be a linear expression greater than a value. Note that all
// coeff must be positive, and we will use the variable lower bound.
std::vector<IntegerVariable> vars;
std::vector<IntegerValue> coeffs;
// Reason for StartMax(before).
const IntegerValue smax_before = StartMax(before);
if (smax_before >= integer_trail_->UpperBound(starts_[before])) {
if (starts_[before].var != kNoIntegerVariable) {
vars.push_back(NegationOf(starts_[before].var));
coeffs.push_back(starts_[before].coeff);
}
} else {
if (ends_[before].var != kNoIntegerVariable) {
vars.push_back(NegationOf(ends_[before].var));
coeffs.push_back(ends_[before].coeff);
}
if (sizes_[before].var != kNoIntegerVariable) {
vars.push_back(sizes_[before].var);
coeffs.push_back(sizes_[before].coeff);
}
}
// Reason for EndMin(after);
const IntegerValue emin_after = EndMin(after);
if (emin_after <= integer_trail_->LowerBound(ends_[after])) {
if (ends_[after].var != kNoIntegerVariable) {
vars.push_back(ends_[after].var);
coeffs.push_back(ends_[after].coeff);
}
} else {
if (starts_[after].var != kNoIntegerVariable) {
vars.push_back(starts_[after].var);
coeffs.push_back(starts_[after].coeff);
}
if (sizes_[after].var != kNoIntegerVariable) {
vars.push_back(sizes_[after].var);
coeffs.push_back(sizes_[after].coeff);
}
}
DCHECK_LT(smax_before, emin_after);
const IntegerValue slack = emin_after - smax_before - 1;
integer_trail_->AppendRelaxedLinearReason(slack, coeffs, vars,
&integer_reason_);
}
bool SchedulingConstraintHelper::PushIntegerLiteral(IntegerLiteral lit) {
CHECK(extra_explanation_callback_ == nullptr);
return integer_trail_->Enqueue(lit, literal_reason_, integer_reason_);
}
bool SchedulingConstraintHelper::PushIntegerLiteralIfTaskPresent(
int t, IntegerLiteral lit) {
if (IsAbsent(t)) return true;
FlagItemAsUsedInReason(t);
RunCallbackIfSet();
if (IsOptional(t)) {
return integer_trail_->ConditionalEnqueue(
PresenceLiteral(t), lit, &literal_reason_, &integer_reason_);
}
return integer_trail_->Enqueue(lit, literal_reason_, integer_reason_);
}
// We also run directly the precedence propagator for this variable so that when
// we push an interval start for example, we have a chance to push its end.
bool SchedulingConstraintHelper::PushIntervalBound(int t, IntegerLiteral lit) {
if (!PushIntegerLiteralIfTaskPresent(t, lit)) return false;
if (IsAbsent(t)) return true;
if (!UpdateCachedValues(t)) return false;
recompute_cache_.Clear(t);
return true;
}
bool SchedulingConstraintHelper::IncreaseStartMin(int t, IntegerValue value) {
if (starts_[t].var == kNoIntegerVariable) {
if (value > starts_[t].constant) return PushTaskAbsence(t);
return true;
}
return PushIntervalBound(t, starts_[t].GreaterOrEqual(value));
}
bool SchedulingConstraintHelper::IncreaseEndMin(int t, IntegerValue value) {
if (ends_[t].var == kNoIntegerVariable) {
if (value > ends_[t].constant) return PushTaskAbsence(t);
return true;
}
return PushIntervalBound(t, ends_[t].GreaterOrEqual(value));
}
bool SchedulingConstraintHelper::DecreaseEndMax(int t, IntegerValue value) {
if (ends_[t].var == kNoIntegerVariable) {
if (value < ends_[t].constant) return PushTaskAbsence(t);
return true;
}
return PushIntervalBound(t, ends_[t].LowerOrEqual(value));
}
bool SchedulingConstraintHelper::PushLiteral(Literal l) {
integer_trail_->EnqueueLiteral(l, literal_reason_, integer_reason_);
return true;
}
bool SchedulingConstraintHelper::PushTaskAbsence(int t) {
if (IsAbsent(t)) return true;
if (!IsOptional(t)) return ReportConflict();
FlagItemAsUsedInReason(t);
if (IsPresent(t)) {
literal_reason_.push_back(Literal(reason_for_presence_[t]).Negated());
return ReportConflict();
}
RunCallbackIfSet();
integer_trail_->EnqueueLiteral(Literal(reason_for_presence_[t]).Negated(),
literal_reason_, integer_reason_);
return true;
}
bool SchedulingConstraintHelper::PushTaskPresence(int t) {
DCHECK_NE(reason_for_presence_[t], kNoLiteralIndex);
DCHECK(!IsPresent(t));
FlagItemAsUsedInReason(t);
if (IsAbsent(t)) {
literal_reason_.push_back(Literal(reason_for_presence_[t]));
return ReportConflict();
}
RunCallbackIfSet();
integer_trail_->EnqueueLiteral(Literal(reason_for_presence_[t]),
literal_reason_, integer_reason_);
return true;
}
bool SchedulingConstraintHelper::PushTaskOrderWhenPresent(int t_before,
int t_after) {
CHECK_NE(t_before, t_after);
if (IsAbsent(t_before)) return true;
if (IsAbsent(t_after)) return true;
if (!IsPresent(t_before) && !IsPresent(t_after)) return true;
const auto [expr, rhs] =
EncodeDifferenceLowerThan(ends_[t_before], starts_[t_after], 0);
const auto status = linear2_bounds_->GetStatus(expr, kMinIntegerValue, rhs);
if (status == RelationStatus::IS_TRUE) return true;
if (status == RelationStatus::IS_FALSE) {
LinearExpression2 negated_expr = expr;
negated_expr.Negate();
linear2_bounds_->AddReasonForUpperBoundLowerThan(
negated_expr, -rhs - 1, &literal_reason_, &integer_reason_);
if (!IsPresent(t_before)) {
AddPresenceReason(t_after);
return PushTaskAbsence(t_before);
} else if (!IsPresent(t_after)) {
AddPresenceReason(t_before);
return PushTaskAbsence(t_after);
} else {
AddPresenceReason(t_before);
AddPresenceReason(t_after);
return ReportConflict();
}
}
if (!IsPresent(t_before) || !IsPresent(t_after)) return true;
AddPresenceReason(t_before);
AddPresenceReason(t_after);
FlagItemAsUsedInReason(t_before);
FlagItemAsUsedInReason(t_after);
RunCallbackIfSet();
return linear2_bounds_->EnqueueLowerOrEqual(expr, rhs, literal_reason_,
integer_reason_);
}
bool SchedulingConstraintHelper::ReportConflict() {
RunCallbackIfSet();
return integer_trail_->ReportConflict(literal_reason_, integer_reason_);
}
void SchedulingConstraintHelper::WatchAllTasks(int id) {
// It is more efficient to enqueue the propagator
// when the helper Propagate() is called. This result in less entries in our
// watched lists.
propagator_ids_.push_back(id);
}
void SchedulingConstraintHelper::FlagItemAsUsedInReason(int t) {
if (extra_explanation_callback_ == nullptr) {
return;
}
used_items_for_reason_.Set(t);
}
void SchedulingConstraintHelper::RunCallbackIfSet() {
if (extra_explanation_callback_ == nullptr) return;
extra_explanation_callback_(used_items_for_reason_.PositionsSetAtLeastOnce(),
&literal_reason_, &integer_reason_);
}
void SchedulingConstraintHelper::ImportReasonsFromOther(
const SchedulingConstraintHelper& other_helper) {
CHECK(other_helper.extra_explanation_callback_ == nullptr);
literal_reason_.insert(literal_reason_.end(),
other_helper.literal_reason_.begin(),
other_helper.literal_reason_.end());
integer_reason_.insert(integer_reason_.end(),
other_helper.integer_reason_.begin(),
other_helper.integer_reason_.end());
}
std::string SchedulingConstraintHelper::TaskDebugString(int t) const {
return absl::StrCat("t=", t, " is_present=",
(IsPresent(t) ? "1"
: IsAbsent(t) ? "0"
: "?"),
" size=[", SizeMin(t).value(), ",", SizeMax(t).value(),
"]", " start=[", StartMin(t).value(), ",",
StartMax(t).value(), "]", " end=[", EndMin(t).value(),
",", EndMax(t).value(), "]");
}
IntegerValue SchedulingConstraintHelper::GetMinOverlap(int t,
IntegerValue start,
IntegerValue end) const {
return std::min(std::min(end - start, SizeMin(t)),
std::min(EndMin(t) - start, end - StartMax(t)));
}
IntegerValue ComputeEnergyMinInWindow(
IntegerValue start_min, IntegerValue start_max, IntegerValue end_min,
IntegerValue end_max, IntegerValue size_min, IntegerValue demand_min,
absl::Span<const LiteralValueValue> filtered_energy,
IntegerValue window_start, IntegerValue window_end) {
if (window_end <= window_start) return IntegerValue(0);
// Returns zero if the interval do not necessarily overlap.
if (end_min <= window_start) return IntegerValue(0);
if (start_max >= window_end) return IntegerValue(0);
const IntegerValue window_size = window_end - window_start;
const IntegerValue simple_energy_min =
demand_min * std::min({end_min - window_start, window_end - start_max,
size_min, window_size});
if (filtered_energy.empty()) return simple_energy_min;
IntegerValue result = kMaxIntegerValue;
for (const auto [lit, fixed_size, fixed_demand] : filtered_energy) {
const IntegerValue alt_end_min = std::max(end_min, start_min + fixed_size);
const IntegerValue alt_start_max =
std::min(start_max, end_max - fixed_size);
const IntegerValue energy_min =
fixed_demand *
std::min({alt_end_min - window_start, window_end - alt_start_max,
fixed_size, window_size});
result = std::min(result, energy_min);
}
if (result == kMaxIntegerValue) return simple_energy_min;
return std::max(simple_energy_min, result);
}
SchedulingDemandHelper::SchedulingDemandHelper(
absl::Span<const AffineExpression> demands,
SchedulingConstraintHelper* helper, Model* model)
: integer_trail_(model->GetOrCreate<IntegerTrail>()),
product_decomposer_(model->GetOrCreate<ProductDecomposer>()),
sat_solver_(model->GetOrCreate<SatSolver>()),
assignment_(model->GetOrCreate<SatSolver>()->Assignment()),
demands_(demands.begin(), demands.end()),
helper_(helper) {
const int num_tasks = helper->NumTasks();
decomposed_energies_.resize(num_tasks);
cached_energies_min_.resize(num_tasks, kMinIntegerValue);
cached_energies_max_.resize(num_tasks, kMaxIntegerValue);
energy_is_quadratic_.resize(num_tasks, false);
// We try to init decomposed energies. This is needed for the cuts that are
// created after we call InitAllDecomposedEnergies().
InitDecomposedEnergies();
}
void SchedulingDemandHelper::InitDecomposedEnergies() {
// For the special case were demands is empty.
const int num_tasks = helper_->NumTasks();
if (demands_.size() != num_tasks) return;
for (int t = 0; t < num_tasks; ++t) {
const AffineExpression size = helper_->Sizes()[t];
const AffineExpression demand = demands_[t];
decomposed_energies_[t] = product_decomposer_->TryToDecompose(size, demand);
}
}
IntegerValue SchedulingDemandHelper::SimpleEnergyMin(int t) const {
if (demands_.empty()) return kMinIntegerValue;
return CapProdI(DemandMin(t), helper_->SizeMin(t));
}
IntegerValue SchedulingDemandHelper::DecomposedEnergyMin(int t) const {
if (decomposed_energies_[t].empty()) return kMinIntegerValue;
IntegerValue result = kMaxIntegerValue;
for (const auto [lit, fixed_size, fixed_demand] : decomposed_energies_[t]) {
if (assignment_.LiteralIsTrue(lit)) {
return fixed_size * fixed_demand;
}
if (assignment_.LiteralIsFalse(lit)) continue;
result = std::min(result, fixed_size * fixed_demand);
}
DCHECK_NE(result, kMaxIntegerValue);
return result;
}
IntegerValue SchedulingDemandHelper::SimpleEnergyMax(int t) const {
if (demands_.empty()) return kMaxIntegerValue;
return CapProdI(DemandMax(t), helper_->SizeMax(t));
}
IntegerValue SchedulingDemandHelper::DecomposedEnergyMax(int t) const {
if (decomposed_energies_[t].empty()) return kMaxIntegerValue;
IntegerValue result = kMinIntegerValue;
for (const auto [lit, fixed_size, fixed_demand] : decomposed_energies_[t]) {
if (assignment_.LiteralIsTrue(lit)) {
return fixed_size * fixed_demand;
}
if (assignment_.LiteralIsFalse(lit)) continue;
result = std::max(result, fixed_size * fixed_demand);
}
DCHECK_NE(result, kMinIntegerValue);
return result;
}
bool SchedulingDemandHelper::CacheAllEnergyValues() {
const int num_tasks = cached_energies_min_.size();
const bool is_at_level_zero = sat_solver_->CurrentDecisionLevel() == 0;
for (int t = 0; t < num_tasks; ++t) {
// Try to reduce the size of the decomposed energy vector.
if (is_at_level_zero) {
int new_size = 0;
for (int i = 0; i < decomposed_energies_[t].size(); ++i) {
if (assignment_.LiteralIsFalse(decomposed_energies_[t][i].literal)) {
continue;
}
decomposed_energies_[t][new_size++] = decomposed_energies_[t][i];
}
decomposed_energies_[t].resize(new_size);
}
cached_energies_min_[t] =
std::max(SimpleEnergyMin(t), DecomposedEnergyMin(t));
if (cached_energies_min_[t] <= kMinIntegerValue) return false;
energy_is_quadratic_[t] =
decomposed_energies_[t].empty() && !demands_.empty() &&
!integer_trail_->IsFixed(demands_[t]) && !helper_->SizeIsFixed(t);
cached_energies_max_[t] =
std::min(SimpleEnergyMax(t), DecomposedEnergyMax(t));
if (cached_energies_max_[t] >= kMaxIntegerValue) return false;
}
return true;
}
IntegerValue SchedulingDemandHelper::DemandMin(int t) const {
DCHECK_LT(t, demands_.size());
return integer_trail_->LowerBound(demands_[t]);
}
IntegerValue SchedulingDemandHelper::DemandMax(int t) const {
DCHECK_LT(t, demands_.size());
return integer_trail_->UpperBound(demands_[t]);
}
bool SchedulingDemandHelper::DemandIsFixed(int t) const {
return integer_trail_->IsFixed(demands_[t]);
}
bool SchedulingDemandHelper::DecreaseEnergyMax(int t, IntegerValue value) {
if (helper_->IsAbsent(t)) return true;
if (value < EnergyMin(t)) return helper_->PushTaskAbsence(t);
if (!decomposed_energies_[t].empty()) {
for (const auto [lit, fixed_size, fixed_demand] : decomposed_energies_[t]) {
if (fixed_size * fixed_demand > value) {
// `lit` encodes that the energy is higher than value. So either lit
// must be false or the task must be absent.
if (assignment_.LiteralIsFalse(lit)) continue;
if (assignment_.LiteralIsTrue(lit)) {
// Task must be absent.
if (helper_->PresenceLiteral(t) != lit) {
helper_->AddLiteralReason(lit.Negated());
}
return helper_->PushTaskAbsence(t);
}
if (helper_->IsPresent(t)) {
// Task is present, `lit` must be false.
DCHECK(!helper_->IsOptional(t) || helper_->PresenceLiteral(t) != lit);
helper_->AddPresenceReason(t);
if (!helper_->PushLiteral(lit.Negated())) return false;
}
}
}
} else {
// TODO(user): Propagate if possible.
VLOG(3) << "Cumulative energy missed propagation";
}
return true;
}
void SchedulingDemandHelper::AddDemandMinReason(int t) {
DCHECK_LT(t, demands_.size());
if (demands_[t].var != kNoIntegerVariable) {
helper_->AddIntegerReason(
integer_trail_->LowerBoundAsLiteral(demands_[t].var));
}
}
void SchedulingDemandHelper::AddDemandMinReason(int t,
IntegerValue min_demand) {
DCHECK_LT(t, demands_.size());
if (demands_[t].var != kNoIntegerVariable) {
helper_->AddIntegerReason(demands_[t].GreaterOrEqual(min_demand));
}
}
void SchedulingDemandHelper::AddEnergyMinReason(int t) {
// We prefer these reason in order.
const IntegerValue value = cached_energies_min_[t];
if (DecomposedEnergyMin(t) >= value) {
for (const auto [lit, fixed_size, fixed_demand] : decomposed_energies_[t]) {
if (assignment_.LiteralIsTrue(lit)) {
helper_->AddLiteralReason(lit.Negated());
return;
}
}
for (const auto [lit, fixed_size, fixed_demand] : decomposed_energies_[t]) {
if (fixed_size * fixed_demand < value &&
assignment_.LiteralIsFalse(lit)) {
helper_->AddLiteralReason(lit);
}
}
} else if (SimpleEnergyMin(t) >= value) {
AddDemandMinReason(t);
helper_->AddSizeMinReason(t);
}
}
bool SchedulingDemandHelper::AddLinearizedDemand(
int t, LinearConstraintBuilder* builder) const {
if (helper_->IsPresent(t)) {
if (!decomposed_energies_[t].empty()) {
for (const LiteralValueValue& entry : decomposed_energies_[t]) {
if (!builder->AddLiteralTerm(entry.literal, entry.right_value)) {
return false;
}
}
} else {
builder->AddTerm(demands_[t], IntegerValue(1));
}
} else if (!helper_->IsAbsent(t)) {
return builder->AddLiteralTerm(helper_->PresenceLiteral(t), DemandMin(t));
}
return true;
}
std::vector<LiteralValueValue> SchedulingDemandHelper::FilteredDecomposedEnergy(
int index) {
if (decomposed_energies_[index].empty()) return {};
if (sat_solver_->CurrentDecisionLevel() == 0) {
// CacheAllEnergyValues has already filtered false literals.
return decomposed_energies_[index];
}
// Scan and filter false literals.
std::vector<LiteralValueValue> result;
for (const auto& e : decomposed_energies_[index]) {
if (assignment_.LiteralIsFalse(e.literal)) continue;
result.push_back(e);
}
return result;
}
void SchedulingDemandHelper::OverrideDecomposedEnergies(
const std::vector<std::vector<LiteralValueValue>>& energies) {
DCHECK_EQ(energies.size(), helper_->NumTasks());
decomposed_energies_ = energies;
}
IntegerValue SchedulingDemandHelper::EnergyMinInWindow(
int t, IntegerValue window_start, IntegerValue window_end) {
return ComputeEnergyMinInWindow(
helper_->StartMin(t), helper_->StartMax(t), helper_->EndMin(t),
helper_->EndMax(t), helper_->SizeMin(t), DemandMin(t),
FilteredDecomposedEnergy(t), window_start, window_end);
}
// Since we usually ask way less often for the reason, we redo the computation
// here.
void SchedulingDemandHelper::AddEnergyMinInWindowReason(
int t, IntegerValue window_start, IntegerValue window_end) {
const IntegerValue actual_energy_min =
EnergyMinInWindow(t, window_start, window_end);
if (actual_energy_min == 0) return;
// Return simple reason right away if there is no decomposition or the simple
// energy is enough.
const IntegerValue start_max = helper_->StartMax(t);
const IntegerValue end_min = helper_->EndMin(t);
const IntegerValue min_overlap =
helper_->GetMinOverlap(t, window_start, window_end);
const IntegerValue simple_energy_min = DemandMin(t) * min_overlap;
if (simple_energy_min == actual_energy_min) {
AddDemandMinReason(t);
helper_->AddSizeMinReason(t);
helper_->AddStartMaxReason(t, start_max);
helper_->AddEndMinReason(t, end_min);
return;
}
// TODO(user): only include the one we need?
const IntegerValue start_min = helper_->StartMin(t);
const IntegerValue end_max = helper_->EndMax(t);
DCHECK(!decomposed_energies_[t].empty());
helper_->AddStartMinReason(t, start_min);
helper_->AddStartMaxReason(t, start_max);
helper_->AddEndMinReason(t, end_min);
helper_->AddEndMaxReason(t, end_max);
DCHECK(!decomposed_energies_[t].empty());
for (const auto [lit, fixed_size, fixed_demand] : decomposed_energies_[t]) {
// Should be the same in most cases.
if (assignment_.LiteralIsTrue(lit)) {
helper_->AddLiteralReason(lit.Negated());
return;
}
}
for (const auto [lit, fixed_size, fixed_demand] : decomposed_energies_[t]) {
if (assignment_.LiteralIsFalse(lit)) {
const IntegerValue alt_em = std::max(end_min, start_min + fixed_size);
const IntegerValue alt_sm = std::min(start_max, end_max - fixed_size);
const IntegerValue energy_min =
fixed_demand *
std::min({alt_em - window_start, window_end - alt_sm, fixed_size});
if (energy_min >= actual_energy_min) continue;
helper_->AddLiteralReason(lit);
}
}
}
void SchedulingConstraintHelper::AddReasonForUpperBoundLowerThan(
LinearExpression2 expr, IntegerValue ub) {
linear2_bounds_->AddReasonForUpperBoundLowerThan(expr, ub, &literal_reason_,
&integer_reason_);
}
void SchedulingConstraintHelper::AppendAndResetReason(
std::vector<IntegerLiteral>* integer_reason,
std::vector<Literal>* literal_reason) {
RunCallbackIfSet();
for (const IntegerLiteral l : integer_reason_) {
integer_reason->push_back(l);
}
for (const Literal l : literal_reason_) {
literal_reason->push_back(l);
}
ResetReason();
}
void AddIntegerVariableFromIntervals(const SchedulingConstraintHelper* helper,
Model* model,
std::vector<IntegerVariable>* vars,
int mask) {
IntegerEncoder* encoder = model->GetOrCreate<IntegerEncoder>();
for (int t = 0; t < helper->NumTasks(); ++t) {
if ((mask & IntegerVariablesToAddMask::kStart) &&
helper->Starts()[t].var != kNoIntegerVariable) {
vars->push_back(helper->Starts()[t].var);
}
if ((mask & IntegerVariablesToAddMask::kSize) &&
helper->Sizes()[t].var != kNoIntegerVariable) {
vars->push_back(helper->Sizes()[t].var);
}
if ((mask & IntegerVariablesToAddMask::kEnd) &&
helper->Ends()[t].var != kNoIntegerVariable) {
vars->push_back(helper->Ends()[t].var);
}
if ((mask & IntegerVariablesToAddMask::kPresence) &&
helper->IsOptional(t) && !helper->IsAbsent(t) &&
!helper->IsPresent(t)) {
const Literal l = helper->PresenceLiteral(t);
IntegerVariable view = kNoIntegerVariable;
if (!encoder->LiteralOrNegationHasView(l, &view)) {
view = model->Add(NewIntegerVariableFromLiteral(l));
}
vars->push_back(view);
}
}
}
void AppendVariablesFromCapacityAndDemands(
const AffineExpression& capacity, SchedulingDemandHelper* demands_helper,
Model* model, std::vector<IntegerVariable>* vars) {
auto* integer_trail = model->GetOrCreate<IntegerTrail>();
for (const AffineExpression& demand_expr : demands_helper->Demands()) {
if (!integer_trail->IsFixed(demand_expr)) {
vars->push_back(demand_expr.var);
}
}
IntegerEncoder* encoder = model->GetOrCreate<IntegerEncoder>();
for (const auto& product : demands_helper->DecomposedEnergies()) {
for (const auto& lit_val_val : product) {
IntegerVariable view = kNoIntegerVariable;
if (!encoder->LiteralOrNegationHasView(lit_val_val.literal, &view)) {
view = model->Add(NewIntegerVariableFromLiteral(lit_val_val.literal));
}
vars->push_back(view);
}
}
if (!integer_trail->IsFixed(capacity)) {
vars->push_back(capacity.var);
}
}
} // namespace sat
} // namespace operations_research