From e89fffdd473dd60ecd1fa522a44a45c9d94cd95e Mon Sep 17 00:00:00 2001 From: Ishika Roy <41401566+Iroy30@users.noreply.github.com> Date: Wed, 4 Feb 2026 14:05:36 -0600 Subject: [PATCH 1/3] Add regression testing (#556) ## Issue Authors: - Ishika Roy (https://github.com/Iroy30) - Ramakrishnap (https://github.com/rgsl888prabhu) Approvers: - Ramakrishnap (https://github.com/rgsl888prabhu) - James Lamb (https://github.com/jameslamb) URL: https://github.com/NVIDIA/cuopt/pull/556 --- python/libcuopt/pyproject.toml | 3 +- regression/benchmark_scripts/benchmark.py | 382 ++++++++ .../benchmark_scripts/configs/README.md | 19 + .../configs/example_test_creation.py | 76 ++ .../configs/test_name_config.json | 23 + .../configs/test_name_data.json | 117 +++ .../benchmark_scripts/results/test_name.csv | 2 + regression/benchmark_scripts/utils.py | 70 ++ regression/config.sh | 60 ++ regression/create-html-reports.sh | 236 +++++ regression/cronjob.sh | 170 ++++ regression/functions.sh | 135 +++ regression/get_datasets.py | 915 ++++++++++++++++++ regression/lp_config.json | 13 + regression/lp_regression_test.sh | 29 + regression/mip_config.json | 29 + regression/mip_regression_test.sh | 28 + regression/routing_regression_test.sh | 29 + regression/run_regression.sh | 16 + regression/save_benchmark_results.py | 52 + regression/save_benchmarks.sh | 38 + regression/send-slack-report.sh | 119 +++ regression/slack_msg.json | 68 ++ regression/test-container.sh | 26 + regression/update_asv_database.py | 157 +++ regression/write-meta-data.sh | 26 + 26 files changed, 2837 insertions(+), 1 deletion(-) create mode 100644 regression/benchmark_scripts/benchmark.py create mode 100644 regression/benchmark_scripts/configs/README.md create mode 100644 regression/benchmark_scripts/configs/example_test_creation.py create mode 100644 regression/benchmark_scripts/configs/test_name_config.json create mode 100644 regression/benchmark_scripts/configs/test_name_data.json create mode 100644 regression/benchmark_scripts/results/test_name.csv create mode 100644 regression/benchmark_scripts/utils.py create mode 100644 regression/config.sh create mode 100755 regression/create-html-reports.sh create mode 100755 regression/cronjob.sh create mode 100644 regression/functions.sh create mode 100644 regression/get_datasets.py create mode 100644 regression/lp_config.json create mode 100644 regression/lp_regression_test.sh create mode 100644 regression/mip_config.json create mode 100644 regression/mip_regression_test.sh create mode 100644 regression/routing_regression_test.sh create mode 100644 regression/run_regression.sh create mode 100644 regression/save_benchmark_results.py create mode 100644 regression/save_benchmarks.sh create mode 100755 regression/send-slack-report.sh create mode 100644 regression/slack_msg.json create mode 100644 regression/test-container.sh create mode 100644 regression/update_asv_database.py create mode 100755 regression/write-meta-data.sh diff --git a/python/libcuopt/pyproject.toml b/python/libcuopt/pyproject.toml index 5286c52529..e2c1142523 100644 --- a/python/libcuopt/pyproject.toml +++ b/python/libcuopt/pyproject.toml @@ -53,7 +53,8 @@ libcuopt = "libcuopt" select = [ "distro-too-large-compressed", ] -max_allowed_size_compressed = '650M' + +max_allowed_size_compressed = '660M' [project.scripts] cuopt_cli = "libcuopt._cli_wrapper:main" diff --git a/regression/benchmark_scripts/benchmark.py b/regression/benchmark_scripts/benchmark.py new file mode 100644 index 0000000000..ead8fe98a4 --- /dev/null +++ b/regression/benchmark_scripts/benchmark.py @@ -0,0 +1,382 @@ +# SPDX-FileCopyrightText: Copyright (c) 2024-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. +# SPDX-License-Identifier: Apache-2.0 + + +import os +from multiprocessing import Process +import rmm +import time +import pandas as pd +import glob +import logging as log +from datetime import datetime +import argparse + +log.getLogger().setLevel(log.INFO) + + +def create_regression_markdown(data, regression_path, test_type_string): + regression_md_file = ( + regression_path + "/" + test_type_string + "_regressions.md" + ) + + md_data = "*No regressions*" + # This to reduce size of slack message + limit_no_of_regression_list = 5 + + if len(data) > 0: + status = "*!! Regressions found !!*" + end_msg = ( + "\n*Continues ...*" + if len(data) > limit_no_of_regression_list + else "" + ) + table = data[:limit_no_of_regression_list].to_string(index=False) + md_data = status + f"\n```\n{table}\n```" + end_msg + + with open(regression_md_file, "w") as fp: + fp.write(md_data) + + +def record_regressions( + test_name, data, req_metrics, regression_path, test_type_string +): + regression_file = ( + regression_path + "/" + test_type_string + "_regressions.csv" + ) + + regression_df = pd.DataFrame( + { + "Test Name": [], + "Metric Name": [], + "Value": [], + "Avg Value": [], + "Regression(%)": [], + } + ) + for name in req_metrics: + if name.startswith("bks_change_"): + pchange = data[name].iloc[-1].item() + metric_name = name.replace("bks_change_", "") + limit = req_metrics[metric_name]["bks"].get("threshold", 5) + prev_val_mean = pchange + latest_val = pchange + else: + limit = req_metrics[name].get("threshold", 5) + prev_val_mean = ( + data[name][:-1][-30:].mean().item() + if len(data) > 1 + else data[name].iloc[-1].item() + ) + latest_val = data[name].iloc[-1].item() + + if prev_val_mean == 0: + pchange = latest_val + else: + pchange = ((latest_val - prev_val_mean) / prev_val_mean) * 100 + + if abs(pchange) >= limit: + regression_df.loc[len(regression_df)] = [ + test_name, + name, + latest_val, + prev_val_mean, + pchange, + ] + + regression_df.to_csv(regression_file) + create_regression_markdown( + regression_df, regression_path, test_type_string + ) + + +def get_bks_change(metrics, required_metrics): + bks_metrics = {} + for metric, value in required_metrics.items(): + if "bks" in value.keys(): + bks = value["bks"]["value"] + if bks is None: + continue + current = metrics[metric] + if bks == 0: + bks_metrics["bks_change_" + metric] = abs(current) * 100 + elif current == 0: + bks_metrics["bks_change_" + metric] = abs(bks) * 100 + else: + bks_metrics["bks_change_" + metric] = abs( + ((current - bks) / bks) * 100 + ) + + return bks_metrics + + +def record_result( + test_name, metrics, required_metrics, csv_path, test_type_string +): + file_path = csv_path + "/" + if test_type_string == "lp" or test_type_string == "mip": + file_path += test_type_string + "_" + test_name + ".csv" + else: + file_path += test_name + ".csv" + bks_metrics = get_bks_change(metrics, required_metrics) + + # Add default metrics to data + required_metrics.update(bks_metrics) + metrics.update(bks_metrics) + req_metrics = list(required_metrics.keys()) + ["date_time", "git_commit"] + + current_data = pd.DataFrame( + {key: [metrics[key]] for key in sorted(req_metrics)} + ) + if os.path.isfile(file_path): + previous_data = pd.read_csv(file_path, index_col=0) + updated_data = pd.concat( + [previous_data, current_data], ignore_index=True + ) + else: + updated_data = current_data + record_regressions( + test_name, updated_data, required_metrics, csv_path, test_type_string + ) + updated_data.to_csv(file_path) + + +def run_benchmark( + test_name, + data_model, + solver_settings, + required_metrics, + csv_path, + git_commit, + test_status_file, + d_type, +): + import rmm + + mr = rmm.mr.get_current_device_resource() + + from utils import LPMetrics, RoutingMetrics + from cuopt import linear_programming + from cuopt import routing + + start_time = time.time() + if d_type == "lp" or d_type == "mip": + metrics = LPMetrics()._asdict() + solver_settings.set_parameter("infeasibility_detection", False) + solver_settings.set_parameter("time_limit", 60) + solution = linear_programming.Solve(data_model, solver_settings) + else: + metrics = RoutingMetrics()._asdict() + solution = routing.Solve(data_model) + end_time = time.time() + + metrics["gpu_memory_usage"] = int( + mr.allocation_counts.peak_bytes / (1024 * 1024) + ) + metrics["date_time"] = datetime.now().strftime("%m_%d_%Y_%H_%M_%S") + metrics["git_commit"] = git_commit + + success_status = False + + if d_type == "lp" or d_type == "mip": + ## Optimal solution + acceptable_termination = ["Optimal", "TimeLimit", "FeasibleFound"] + if solution.get_termination_reason() in acceptable_termination: + success_status = True + metrics["solver_time"] = solution.get_solve_time() + metrics["primal_objective_value"] = solution.get_primal_objective() + if d_type == "lp": + lp_stats = solution.get_lp_stats() + metrics["nb_iterations"] = lp_stats["nb_iterations"] + else: + milp_stats = solution.get_milp_stats() + metrics["mip_gap"] = milp_stats["mip_gap"] + metrics["max_constraint_violation"] = milp_stats[ + "max_constraint_violation" + ] + metrics["max_int_violation"] = milp_stats["max_int_violation"] + metrics["max_variable_bound_violation"] = milp_stats[ + "max_variable_bound_violation" + ] + record_result( + test_name, metrics, required_metrics, csv_path, d_type + ) + else: + if solution.get_status() == 0: + success_status = True + metrics["solver_time"] = end_time - start_time + metrics["total_objective_value"] = solution.get_total_objective() + metrics["vehicle_count"] = solution.get_vehicle_count() + + objectives = solution.get_objective_values() + if "prize" in required_metrics: + metrics["prize"] = objectives[routing.Objective.PRIZE] + if "cost" in required_metrics: + metrics["cost"] = objectives[routing.Objective.COST] + if "travel_time" in required_metrics: + metrics["travel_time"] = objectives[ + routing.Objective.TRAVEL_TIME + ] + record_result( + test_name, metrics, required_metrics, csv_path, d_type + ) + return "SUCCESS" if success_status is True else "FAILED" + + +def reinitialize_rmm(): + pool_size = 2**30 + rmm.reinitialize(pool_allocator=True, initial_pool_size=pool_size) + + base_mr = rmm.mr.get_current_device_resource() + stats_mr = rmm.mr.StatisticsResourceAdaptor(base_mr) + rmm.mr.set_current_device_resource(stats_mr) + + return base_mr, stats_mr + + +def worker( + gpu_id, + dataset_file_path, + csv_path, + git_commit, + log_path, + test_status_file, + n_gpus, + d_type="routing", +): + import os + + os.environ["CUDA_VISIBLE_DEVICES"] = gpu_id + + from utils import get_configuration + + data_files = [] + if d_type == "lp" or d_type == "mip": + data_files = glob.glob(dataset_file_path + "/*.mps") + else: + data_files = glob.glob(dataset_file_path + "/*_config.json") + idx = int(gpu_id) + n_files = 1 # len(data_files) + + while idx < n_files: + mr, stats_mr = reinitialize_rmm() + + data_file = data_files[idx] + test_name = str(data_file) + status = "FAILED" + try: + test_name, data_model, solver_settings, requested_metrics = ( + get_configuration(data_file, dataset_file_path, d_type) + ) + log.basicConfig( + level=log.INFO, + filename=log_path + "/" + test_name + "_log.txt", + filemode="a+", + format="%(asctime)-15s %(levelname)-8s %(message)s", + ) + log.getLogger().setLevel(log.INFO) + log.info( + f"------------- Test Start : {test_name} gpu id : {gpu_id} -------------------" + ) + status = run_benchmark( + test_name, + data_model, + solver_settings, + requested_metrics, + csv_path, + git_commit, + test_status_file, + d_type, + ) + except Exception as e: + log.error(str(e)) + + with open(test_status_file, "a") as f: + f.write("\n") + f.write(test_name + ": " + status) + + # Delete instance of rmm + del mr + del stats_mr + + log.info( + f"------------- Test End : {test_name} gpu id : {gpu_id} -------------------" + ) + idx = idx + n_gpus + + +def run( + dataset_file_path, + csv_path, + git_commit, + log_path, + test_status_file, + n_gpus, + d_type, +): + # Restricting n_gpus to one to avoid resource sharing + # n_gpus = 1 + procs = [] + for gpu_id in range(int(n_gpus)): + p = Process( + target=worker, + args=( + str(gpu_id), + dataset_file_path, + csv_path, + git_commit, + log_path, + test_status_file, + int(n_gpus), + d_type, + ), + ) + p.start() + procs.append(p) + + for p in procs: + p.join() + print("All processes finished.") + + +if __name__ == "__main__": + parser = argparse.ArgumentParser() + + parser.add_argument( + "-c", "--config-path", type=str, help="Path to all configuration file" + ) + parser.add_argument( + "-r", + "--csv-path", + type=str, + help="Path to store result files, this would be for github where results gets stored", + ) + parser.add_argument( + "-g", + "--git-commit", + type=str, + help="git commit sha to keep track of runs", + ) + parser.add_argument("-l", "--log-path", type=str, help="Path to log files") + parser.add_argument( + "-s", + "--test-status-file", + type=str, + help="All test status will be stored in this file", + ) + parser.add_argument( + "-n", "--num-gpus", type=str, help="Number of GPUs available" + ) + parser.add_argument( + "-t", "--type", type=str, default="", help="Type of benchmark" + ) + args = parser.parse_args() + run( + args.config_path, + args.csv_path, + args.git_commit, + args.log_path, + args.test_status_file, + args.num_gpus, + args.type, + ) diff --git a/regression/benchmark_scripts/configs/README.md b/regression/benchmark_scripts/configs/README.md new file mode 100644 index 0000000000..fb82836563 --- /dev/null +++ b/regression/benchmark_scripts/configs/README.md @@ -0,0 +1,19 @@ +# Creating configuration and data file for routing + +- For each test, create a configuration file and a corresponding data file. +- Refer `test_name_config.json` for the format of the configuration file. +- Supported metrics can be found in `cuopt/regression/benchmark_scripts/utils.py` +- File names should start with test names followed by `config` or data depending on type of it. +- Data file should be as per openapi spec of cuopt server +- These configuration and data files needs to be uploaded to `s3://cuopt-datasets/regression_datasets/` + + ``` + aws s3 cp /path/to/files s3://cuopt-datasets/regression_datasets/ + ``` + +# Creating configuration and data file for lp and milp + +- For each test, create a mps file +- Refer `lp_config.json` and `mip_config.json` for the format of the configuration file. +- Supported metrics can be found in `cuopt/regression/benchmark_scripts/utils.py` +- These configuration and data files needs to be in the LP_DATASETS_PATH set in config.sh diff --git a/regression/benchmark_scripts/configs/example_test_creation.py b/regression/benchmark_scripts/configs/example_test_creation.py new file mode 100644 index 0000000000..0b376eb001 --- /dev/null +++ b/regression/benchmark_scripts/configs/example_test_creation.py @@ -0,0 +1,76 @@ +# SPDX-FileCopyrightText: Copyright (c) 2024-2026, NVIDIA CORPORATION. +# SPDX-License-Identifier: Apache-2.0 +# CONFIDENTIAL, provided under NDA. + +from cuopt.routing import utils +import json + +""" +This is an example of creating a modified test from Homberger dataset. +In this test, the RC2_10_5 test is modified so that the vehicle count is reduced to 12 and the order prizes are set. +The prizes are high enough so that prize always becomes the primary objective. +One can easily use an existing json file and modify the data as well by loading the json as a dictionary +""" +test_name = "prize_collection_vrp" +# test_name = "LC1_10_9" + +# base_file_name = "/home/nfs/rgandham/git-repos/reopt/datasets/pdptw/LC1_10_9.pdptw" +base_file_name = ( + "/home/nfs/rgandham/git-repos/reopt/datasets/cvrptw/RC2_10_5.TXT" +) + +# model_dict = utils.create_model_dictionary_from_file(base_file_name, is_pdp=True) +model_dict = utils.create_model_dictionary_from_file(base_file_name) + + +# Reduce the fleet size to 12 +num_vehicles = 12 +fleet_data = model_dict["fleet_data"] + +vehicle_locations = fleet_data["vehicle_locations"] +vehicle_tw = fleet_data["vehicle_time_windows"] +capacities = fleet_data["capacities"] + +new_locs = [vehicle_locations[i] for i in range(num_vehicles)] +new_tw = [vehicle_tw[i] for i in range(num_vehicles)] +new_cap = [[capacities[0][i] for i in range(num_vehicles)]] * 1 + +fleet_data["vehicle_locations"] = new_locs +fleet_data["vehicle_time_windows"] = new_tw +fleet_data["capacities"] = new_cap + +# Add prizes +task_data = model_dict["task_data"] + +n_tasks = len(task_data["demand"][0]) + +prizes = [10000.0] * n_tasks +task_data["prizes"] = prizes + + +# Set 10 min time limit +solver_config = {} +solver_config["time_limit"] = 600 + +model_dict["solver_config"] = solver_config + +test_config_file_name = test_name + "_config.json" +model_data_file_name = test_name + "_data.json" + +test_config = {} +test_config["test_name"] = test_name +test_config["file_name"] = model_data_file_name +test_config["metrics"] = [ + "vehicle_count", + "total_cost", + "prize", + "memory_usage", +] + +with open(test_config_file_name, "w") as fp: + json.dump(test_config, fp) + fp.close() + +with open(model_data_file_name, "w") as fp: + json.dump(model_dict, fp) + fp.close() diff --git a/regression/benchmark_scripts/configs/test_name_config.json b/regression/benchmark_scripts/configs/test_name_config.json new file mode 100644 index 0000000000..d87b7bc842 --- /dev/null +++ b/regression/benchmark_scripts/configs/test_name_config.json @@ -0,0 +1,23 @@ +{ + "test_name": "test_name", + "file_name": "test_name_data.json", + "metrics": { + "total_objective_value": { + "threshold": 5, + "unit": "total_objective_value" + }, + "vehicle_count": { + "threshold": 5, + "unit": "vehicle_count" + }, + "solver_time": { + "threshold": 5, + "unit": "seconds" + }, + "gpu_memory_usage": { + "threshold": 20, + "unit": "MB" + } + }, + "details": "Add details about you test" +} diff --git a/regression/benchmark_scripts/configs/test_name_data.json b/regression/benchmark_scripts/configs/test_name_data.json new file mode 100644 index 0000000000..e6918ad58e --- /dev/null +++ b/regression/benchmark_scripts/configs/test_name_data.json @@ -0,0 +1,117 @@ +{ + "cost_waypoint_graph_data": { + "waypoint_graph": null + }, + "travel_time_waypoint_graph_data": { + "waypoint_graph": null + }, + "cost_matrix_data": { + "data": { + "0": [ + [ + 0, + 1, + 1 + ], + [ + 1, + 0, + 1 + ], + [ + 1, + 1, + 0 + ] + ] + } + }, + "travel_time_matrix_data": { + "data": null + }, + "task_data": { + "task_locations": [ + 0, + 1, + 2 + ], + "demand": [ + [ + 0, + 1, + 1 + ], + [ + 0, + 3, + 1 + ] + ], + "task_time_windows": [ + [ + 0, + 10 + ], + [ + 0, + 4 + ], + [ + 2, + 4 + ] + ], + "service_times": [ + 0, + 1, + 1 + ] + }, + "fleet_data": { + "vehicle_locations": [ + [ + 0, + 0 + ], + [ + 0, + 0 + ] + ], + "capacities": [ + [ + 2, + 2 + ], + [ + 4, + 1 + ] + ], + "vehicle_time_windows": [ + [ + 0, + 10 + ], + [ + 0, + 10 + ] + ], + "skip_first_trips": [ + false, + false + ], + "drop_return_trips": [ + false, + false + ], + "vehicle_max_costs": [ + 20, + 20 + ] + }, + "solver_config": { + "time_limit": 10 + } +} diff --git a/regression/benchmark_scripts/results/test_name.csv b/regression/benchmark_scripts/results/test_name.csv new file mode 100644 index 0000000000..85bf3d9761 --- /dev/null +++ b/regression/benchmark_scripts/results/test_name.csv @@ -0,0 +1,2 @@ +,solver_time,total_objective_value,vehicle_count +0,10.004132270812988,3.0,1 diff --git a/regression/benchmark_scripts/utils.py b/regression/benchmark_scripts/utils.py new file mode 100644 index 0000000000..04b1cbb9fb --- /dev/null +++ b/regression/benchmark_scripts/utils.py @@ -0,0 +1,70 @@ +# SPDX-FileCopyrightText: Copyright (c) 2024-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. +# SPDX-License-Identifier: Apache-2.0 + + +from cuopt_server.utils.utils import build_routing_datamodel_from_json +from cuopt.linear_programming.solver_settings import SolverSettings +import cuopt_mps_parser +import os +import json +from typing import NamedTuple + + +def build_datamodel_from_mps(data): + """ + data: A file in mps format + """ + + if os.path.isfile(data): + data_model = cuopt_mps_parser.ParseMps(data) + else: + raise ValueError( + f"Invalid type : {type(data)} has been provided as input, " + "requires mps input" + ) + solver_settings = SolverSettings() + + return data_model, solver_settings + + +class RoutingMetrics(NamedTuple): + total_objective_value: float = -1 + vehicle_count: int = -1 + cost: float = -1 + prize: float = -1 + travel_time: float = -1 + solver_time: float = -1 + gpu_memory_usage: float = -1 + git_commit: str = "" + date_time: str = "" + + +class LPMetrics(NamedTuple): + primal_objective_value: float = -1 + solver_time: float = -1 + gpu_memory_usage: float = -1 + git_commit: str = "" + date_time: str = "" + + +def get_configuration(data_file, data_file_path, d_type): + data = {} + test_name = None + requested_metrics = {} + + if d_type == "lp" or d_type == "mip": + with open(data_file_path + "/" + d_type + "_config.json") as f: + data = json.load(f) + test_name = data_file.split("/")[-1].split(".")[0] + data_model, solver_settings = build_datamodel_from_mps(data_file) + requested_metrics = data["metrics"] + else: + with open(data_file) as f: + data = json.load(f) + test_name = data["test_name"] + data_model, solver_settings = build_routing_datamodel_from_json( + data_file_path + "/" + data["file_name"] + ) + requested_metrics = data["metrics"] + + return test_name, data_model, solver_settings, requested_metrics diff --git a/regression/config.sh b/regression/config.sh new file mode 100644 index 0000000000..2b20597d07 --- /dev/null +++ b/regression/config.sh @@ -0,0 +1,60 @@ +#!/bin/bash +# shellcheck disable=all +# SPDX-FileCopyrightText: Copyright (c) 2025-2026, NVIDIA CORPORATION. All rights reserved. +# SPDX-License-Identifier: Apache-2.0 + + +THIS_DIR=$(cd $(dirname ${BASH_SOURCE[0]}) && pwd) + +CUOPT_SCRIPTS_DIR=$THIS_DIR +OUTPUT_DIR=$SCRATCH_DIR/benchmark_runs/ + +ACCOUNT=datascience_rapids_testing +PARTITION="batch" +GPUS_PER_NODE=8 + +# Path to the squashs file containing the container image +IMAGE="nvidia/cuopt:26.2.0a-cuda12.9-py3.13" + +ALL_CONFIGS_PATH=$SCRATCH_DIR/configs/ +ROUTING_CONFIGS_PATH=$SCRATCH_DIR/routing_configs/ +ROUTING_DATASETS_PATH=$SCRATCH_DIR/routing_datasets/ +LP_DATASETS_PATH=$SCRATCH_DIR/lp_datasets/ +MIP_DATASETS_PATH=$SCRATCH_DIR/mip_datasets/ + +STATUS_FILE=$OUTPUT_DIR/status.txt +WORKER_RMM_POOL_SIZE=${WORKER_RMM_POOL_SIZE:-24G} + +DATASETS_DIR=$SCRATCH_DIR/datasets + +# Assume CUOPT_SLACK_APP_ID is defined! +CUOPT_SLACK_APP_ID="MY_SLACK_APP_ID" +WEBHOOK_URL=${WEBHOOK_URL:-https://hooks.slack.com/services/${CUOPT_SLACK_APP_ID}} +S3_FILE_PREFIX="MY_S3_FILE_PREFIX" +S3_URL_PREFIX="MY_S3_URL_PREFIX" + +# Most are defined using the bash := or :- syntax, which means they +# will be set only if they were previously unset. The project config +# is loaded first, which gives it the opportunity to override anything +# in this file that uses that syntax. If there are variables in this +# file that should not be overridded by a project, then they will +# simply not use that syntax and override, since these variables are +# read last. +RESULTS_ARCHIVE_DIR=$OUTPUT_DIR/results +RESULTS_DIR=$RESULTS_ARCHIVE_DIR/latest +METADATA_FILE=$RESULTS_DIR/metadata.sh +WORKSPACE=$OUTPUT_DIR/workspace +TESTING_DIR=$WORKSPACE/testing +BENCHMARK_DIR=$WORKSPACE/benchmark +SCRIPTS_DIR=$THIS_DIR + +BUILD_LOG_FILE=$RESULTS_DIR/build_log.txt +DATE=${DATE:-$(date --utc "+%Y-%m-%d_%H:%M:%S")_UTC} + +# vars that are not overridden by the project. + +# These must remain relative to $RESULTS_DIR since some scripts assume +# that, and also assume the names "tests" and "benchmarks", and +# therefore cannot be overridden by a project. +TESTING_RESULTS_DIR=${RESULTS_DIR}/tests +BENCHMARK_RESULTS_DIR=${RESULTS_DIR}/benchmarks diff --git a/regression/create-html-reports.sh b/regression/create-html-reports.sh new file mode 100755 index 0000000000..5b0883dda8 --- /dev/null +++ b/regression/create-html-reports.sh @@ -0,0 +1,236 @@ +#!/bin/bash +# shellcheck disable=SC1090 +# SPDX-FileCopyrightText: Copyright (c) 2021-2026, NVIDIA CORPORATION. +# SPDX-License-Identifier: Apache-2.0 + +# Must ensure PROJECT_DIR is exported first then load env +export PROJECT_DIR=${PROJECT_DIR:-$(cd "$(dirname ${BASH_SOURCE[0]})" && pwd)} +source ${PROJECT_DIR}/config.sh +source ${PROJECT_DIR}/functions.sh + +# FIXME: this assumes all reports are from running pytests +ALL_REPORTS=$(find ${RESULTS_DIR}/benchmarks/results/ -name "*status.txt") + +# Create the html describing the build and test run +REPORT_METADATA_HTML="" +PROJECT_VERSION="unknown" +PROJECT_BUILD="" +PROJECT_CHANNEL="unknown" +PROJECT_REPO_URL="unknown" +PROJECT_REPO_BRANCH="unknown" +if [ -f $METADATA_FILE ]; then + source $METADATA_FILE +fi +# Assume if PROJECT_BUILD is set then a conda version string should be +# created, else a git version string. +if [[ "$PROJECT_BUILD" != "" ]]; then + REPORT_METADATA_HTML=" + + + +
conda version$PROJECT_VERSION
build$PROJECT_BUILD
channel$PROJECT_CHANNEL
+
" +else + REPORT_METADATA_HTML=" + + + +
commit hash$PROJECT_VERSION
repo$PROJECT_REPO_URL
branch$PROJECT_REPO_BRANCH
+
" +fi + +################################################################################ +# create the html reports for each individual run (each +# pytest-results*.txt file) +if [ "$ALL_REPORTS" != "" ]; then + for report in $ALL_REPORTS; do + # Get the individual report name, and use the .txt file path + # to form the html report being generated (same location as + # the .txt file). This will be an abs path since it is a file + # on disk being written. + report_name=$(basename -s .txt $report) + html_report_abs_path=$(dirname $report)/${report_name}.html + echo " + + + ${report_name} + + +

${report_name}


" > $html_report_abs_path + echo "$REPORT_METADATA_HTML" >> $html_report_abs_path + echo " + + + +" >> $html_report_abs_path + awk '{ if($2 == "FAILED") { + color = "red" + } else { + color = "green" + } + printf "\n", $1, color, $2, $3, $3 + }' $report >> $html_report_abs_path + echo "
test filestatuslogs
%s%s%s
+ + + " >> $html_report_abs_path + done +fi + +################################################################################ +# Create a .html file for each *_log.txt file, which is just the contents +# of the log with a line number and anchor id for each line that can +# be used for sharing links to lines. +ALL_LOGS=$(find -L ${BENCHMARK_RESULTS_DIR} -type f -name "*_log.txt" -print) + +for f in $ALL_LOGS; do + base_no_extension=$(basename ${f: 0:-4}) + html=${f: 0:-4}.html + echo " + + + $base_no_extension + + + +

${base_no_extension}


+" > $html + awk '{ print ""NR":
"$0"

"}' $f >> $html + echo " + +" >> $html +done + +################################################################################ +# create the top-level report +STATUS='FAILED' +STATUS_IMG='https://img.icons8.com/cotton/80/000000/cancel--v1.png' +if [ "$ALL_REPORTS" != "" ]; then + if ! (grep -w FAILED $ALL_REPORTS > /dev/null); then + STATUS='PASSED' + STATUS_IMG='https://img.icons8.com/bubbles/100/000000/approval.png' + fi +fi +BUILD_LOG_HTML="(build log not available or build not run)" +BUILD_STATUS="" +if [ -f $BUILD_LOG_FILE ]; then + if [ -f ${BUILD_LOG_FILE: 0:-4}.html ]; then + BUILD_LOG_HTML="log (plain text)" + else + BUILD_LOG_HTML="log" + fi + tail -3 $BUILD_LOG_FILE | grep -w "done." + if (tail -3 $BUILD_LOG_FILE | grep -qw "done."); then + BUILD_STATUS="PASSED" + else + BUILD_STATUS="FAILED" + fi +fi + +report=${RESULTS_DIR}/report.html +echo " + + + test report + + +" > $report +echo "$REPORT_METADATA_HTML" >> $report +echo "\"${STATUS}\"/ Overall status: $STATUS
" >> $report +echo "Build: ${BUILD_STATUS} ${BUILD_LOG_HTML}
" >> $report +if [ "$ALL_REPORTS" != "" ]; then + echo "
Test Status
" >>$report + echo " + + + + " >> $report + for f in $ALL_REPORTS; do + report_name=$(basename -s .txt $f) + # report_path should be of the form "tests/foo.html" + prefix_to_remove="$RESULTS_DIR/" + report_rel_path=${f/$prefix_to_remove} + report_path=$(dirname $report_rel_path)/${report_name}.html + + if (grep -w FAILED $f > /dev/null); then + status="FAILED" + color="red" + else + status="PASSED" + color="green" + fi + echo "" >> $report + done + echo "
TestStatus
${report_name}${status}
" >> $report +else + echo "Tests were not run." >> $report +fi +prefix_to_remove="$RESULTS_DIR/" +plot_rel_path=${f/$prefix_to_remove} +plot_path=$(dirname $plot_rel_path)/asv/html/index.html +prefix_to_remove="$RESULTS_DIR/benchmarks/results/" +log_rel_path=${f/$prefix_to_remove} +log_path=$(dirname $log_rel_path)/index.html +echo "

\"Plots\"
Plots : Rgression test results


" >>$report +echo "

\"Plots\"
Logs and Details : All the data for this run


" >>$report +echo " + +" >> $report + +################################################################################ +# (optional) generate the ASV html +if hasArg --run-asv; then + asv_config_file=$(find ${BENCHMARK_RESULTS_DIR}/results/asv -name "asv.conf.json") + if [ "$asv_config_file" != "" ]; then + asv update --config $asv_config_file + asv publish --config $asv_config_file + fi +fi + +################################################################################ +# Create an index.html for each dir (ALL_DIRS plus ".", but EXCLUDE +# the asv html) This is needed since S3 (and probably others) will not +# show the contents of a hosted directory by default, but will instead +# return the index.html if present. +# The index.html will just contain links to the individual files and +# subdirs present in each dir, just as if browsing in a file explorer. +ALL_DIRS=$(find -L ${RESULTS_DIR} -path ${BENCHMARK_RESULTS_DIR}/results/asv/html -prune -o -type d -printf "%P\n") + +for d in "." $ALL_DIRS; do + index=${RESULTS_DIR}/${d}/index.html + echo " + + + $d + + +

${d}


+" > $index + for f in ${RESULTS_DIR}/$d/*; do + b=$(basename $f) + # Do not include index.html in index.html (it's a link to itself) + if [[ "$b" == "index.html" ]]; then + continue + fi + if [ -d "$f" ]; then + echo "$b
" >> $index + # special case: if the file is a *_log.txt and has a corresponding .html + elif [[ "${f: -8}" == "_log.txt" ]] && [[ -f "${f: 0:-4}.html" ]]; then + markup="${b: 0:-4}.html" + plaintext=$b + echo "$markup (plain text)
" >> $index + elif [[ "${f: -9}" == "_log.html" ]] && [[ -f "${f: 0:-5}.txt" ]]; then + continue + else + echo "$b
" >> $index + fi + done + echo " + +" >> $index +done diff --git a/regression/cronjob.sh b/regression/cronjob.sh new file mode 100755 index 0000000000..de05b22bc0 --- /dev/null +++ b/regression/cronjob.sh @@ -0,0 +1,170 @@ +#!/bin/bash +# shellcheck disable=SC1090 +# SPDX-FileCopyrightText: Copyright (c) 2021-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. +# SPDX-License-Identifier: Apache-2.0 + +# Abort script on first error to ensure script-env.sh is sourced. +set -e + +if [[ -v SLURM_NODEID ]]; then + echo "Detected the env var SLURM_NODEID is set. Is this script running on a compute node?" + echo "This script must be run *outside* of a slurm job (this script starts slurm jobs, but is not a job itself)." + exit 1 +fi + +# Must ensure PROJECT_DIR is exported first then load rapids-mg-tools env +export PROJECT_DIR=${PROJECT_DIR:-$(cd "$(dirname ${BASH_SOURCE[0]})" && pwd)} + +source ${PROJECT_DIR}/config.sh +source ${PROJECT_DIR}/functions.sh + +RUN_BENCHMARKS=0 + +if hasArg --benchmark; then + RUN_BENCHMARKS=1 +fi + +if (! hasArg --test) && (! hasArg --benchmark); then + RUN_BENCHMARKS=1 +fi + +################################################################################ + +# Create a results dir unique for this run +setupResultsDir + +# Switch to allowing errors from commands, since test failures will +# result in non-zero return codes and this script should attempt to +# run all tests. +set +e + +################################################################################ +logger "Testing cuOpt in container..." +srun \ + --account $ACCOUNT \ + --partition $PARTITION \ + --job-name=test-container.testing \ + --nodes 1 \ + --gpus-per-node 1 \ + --time=120 \ + --export=ALL \ + --container-mounts=${CUOPT_SCRIPTS_DIR}:${CUOPT_SCRIPTS_DIR},${OUTPUT_DIR}:${OUTPUT_DIR} \ + --container-image=$IMAGE \ + --output=$BUILD_LOG_FILE \ + bash ${PROJECT_DIR}/test-container.sh +TESTING_FAILED=$? +logger "done testing container, return code was $TESTING_FAILED" + + +if [[ $TESTING_FAILED == 0 ]]; then + + ############################################################################ + # Setup and run tests + if [[ $RUN_BENCHMARKS == 1 ]]; then + logger "Running benchmarks..." + logger "GPUs per node : $GPUS_PER_NODE" + # SNMG tests - run in parallel + srun \ + --account $ACCOUNT \ + --partition $PARTITION \ + --job-name=run-nightly-benchmarks \ + --nodes 1 \ + --gpus-per-node $GPUS_PER_NODE \ + --time=4:00:00 \ + --export=ALL \ + --exclusive \ + -K \ + --container-mounts ${ROUTING_CONFIGS_PATH}:${ROUTING_CONFIGS_PATH},${CUOPT_SCRIPTS_DIR}:${CUOPT_SCRIPTS_DIR},${OUTPUT_DIR}:${OUTPUT_DIR} \ + --container-image=$IMAGE \ + --output=${BENCHMARK_RESULTS_DIR}/benchmark_routing_log.txt \ + bash ${CUOPT_SCRIPTS_DIR}/routing_regression_test.sh & + PID_1=$! + logger "Process ID $PID_1 in background" + + srun \ + --account $ACCOUNT \ + --partition $PARTITION \ + --job-name=run-nightly-benchmarks \ + --nodes 1 \ + --gpus-per-node $GPUS_PER_NODE \ + --time=4:00:00 \ + --export=ALL \ + --exclusive \ + -K \ + --container-mounts ${LP_DATASETS_PATH}:${LP_DATASETS_PATH},${CUOPT_SCRIPTS_DIR}:${CUOPT_SCRIPTS_DIR},${OUTPUT_DIR}:${OUTPUT_DIR} \ + --container-image=$IMAGE \ + --output=${BENCHMARK_RESULTS_DIR}/benchmark_lp_log.txt \ + bash ${CUOPT_SCRIPTS_DIR}/lp_regression_test.sh & + PID_2=$! + + srun \ + --account $ACCOUNT \ + --partition $PARTITION \ + --job-name=run-nightly-benchmarks \ + --nodes 1 \ + --gpus-per-node $GPUS_PER_NODE \ + --time=4:00:00 \ + --export=ALL \ + --exclusive \ + -K \ + --container-mounts ${MIP_DATASETS_PATH}:${MIP_DATASETS_PATH},${CUOPT_SCRIPTS_DIR}:${CUOPT_SCRIPTS_DIR},${OUTPUT_DIR}:${OUTPUT_DIR} \ + --container-image=$IMAGE \ + --output=${BENCHMARK_RESULTS_DIR}/benchmark_mip_log.txt \ + bash ${CUOPT_SCRIPTS_DIR}/mip_regression_test.sh & + PID_3=$! + + wait $PID_1 $PID_2 $PID_3 + fi + +else # if [[ $TESTING_FAILED == 0 ]] + logger "Container testing Failed!" +fi + +################################################################################ +# Send report based on contents of $RESULTS_DIR +# These steps do not require a worker node. + +# When running both testing and benchmark and if some benchmarks fail, +# the entire nightly will fail. The benchmark logs reported on Slack +# contains information about the failures. +logger "Generating report" + +if [ -f $METADATA_FILE ]; then + source $METADATA_FILE +fi + +# Copy all config files to one folder +cp $ROUTING_CONFIGS_PATH/*config.json $LP_DATASETS_PATH/*config.json $MIP_DATASETS_PATH/*config.json $ALL_CONFIGS_PATH/ + +RUN_ASV_OPTION="" +if hasArg --skip-asv; then + logger "Skipping running ASV" +else + # Only create/update the asv database if there is both a commit Hash and a branch otherwise + # asv will return an error. If there is $PROJECT_BUILD, that implies there is Neither the + # git commit hash nor the branch which are required to create/update the asv db + if [[ "$PROJECT_BUILD" == "" ]]; then + # Update/create the ASV database + logger "Updating ASV database" + python $PROJECT_DIR/update_asv_database.py --commitHash=$PROJECT_VERSION --repo-url=$PROJECT_REPO_URL --branch=$PROJECT_REPO_BRANCH --commitTime=$PROJECT_REPO_TIME --results-dir=$RESULTS_DIR --machine-name=$MACHINE --gpu-type=$GPU_TYPE --configs=$ALL_CONFIGS_PATH + RUN_ASV_OPTION=--run-asv + logger "Updated ASV database" + else + logger "Detected a conda install, cannot run ASV since a commit hash/time is needed." + fi +fi + +# The cuopt pull has missing .git folder which causes subsequent runs, lets delete and pull it fresh everytime. +rm -rf $RESULTS_DIR/benchmarks/results/asv/cuopt/ +rm -rf $RESULTS_DIR/tests + +${SCRIPTS_DIR}/create-html-reports.sh $RUN_ASV_OPTION + +if hasArg --skip-sending-report; then + logger "Skipping sending report." +else + logger "Uploading to S3, posting to Slack" + ${PROJECT_DIR}/send-slack-report.sh +fi + +logger "cronjob.sh done." diff --git a/regression/functions.sh b/regression/functions.sh new file mode 100644 index 0000000000..9d8147e821 --- /dev/null +++ b/regression/functions.sh @@ -0,0 +1,135 @@ +#!/bin/bash +# SPDX-FileCopyrightText: Copyright (c) 2021-2026, NVIDIA CORPORATION. +# SPDX-License-Identifier: Apache-2.0 + +# This file is source'd from script-env.sh to add functions to the +# calling environment, hence no #!/bin/bash as the first line. This +# also assumes the variables used in this file have been defined +# elsewhere. + +NUMARGS=$# +ARGS=$* +function hasArg { + (( ${NUMARGS} != 0 )) && (echo " ${ARGS} " | grep -q " $1 ") +} + +function logger { + echo -e ">>>> $*" +} + +# Calling "setTee outfile" will cause all stdout and stderr of the +# current script to be output to "tee", which outputs to stdout and +# "outfile" simultaneously. This is useful by allowing a script to +# "tee" itself at any point without being called with tee. +_origFileDescriptorsSaved=0 +function setTee { + if [[ $_origFileDescriptorsSaved == 0 ]]; then + # Save off the original file descr 1 and 2 as 3 and 4 + exec 3>&1 4>&2 + _origFileDescriptorsSaved=1 + fi + teeFile=$1 + # Create a named pipe. + pipeName=$(mktemp -u) + mkfifo $pipeName + # Close the currnet 1 and 2 and restore to original (3, 4) in the + # event this function is called repeatedly. + exec 1>&- 2>&- + exec 1>&3 2>&4 + # Start a tee process reading from the named pipe. Redirect stdout + # and stderr to the named pipe which goes to the tee process. The + # named pipe "file" can be removed and the tee process stays alive + # until the fd is closed. + tee -a < $pipeName $teeFile & + exec > $pipeName 2>&1 + rm $pipeName +} + +# Call this to stop script output from going to "tee" after a prior +# call to setTee. +function unsetTee { + if [[ $_origFileDescriptorsSaved == 1 ]]; then + # Close the current fd 1 and 2 which should stop the tee + # process, then restore 1 and 2 to original (saved as 3, 4). + exec 1>&- 2>&- + exec 1>&3 2>&4 + fi +} + +# Creates a unique results dir based on date, then links the common +# results dir name to it. +function setupResultsDir { + mkdir -p ${RESULTS_ARCHIVE_DIR}/${DATE} + # Store the target of $RESULTS_DIR before $RESULTS_DIR get linked to + # a different dir + previous_results=$(readlink -f $RESULTS_DIR) + + rm -rf $RESULTS_DIR + ln -s ${RESULTS_ARCHIVE_DIR}/${DATE} $RESULTS_DIR + mkdir -p $TESTING_RESULTS_DIR + mkdir -p $BENCHMARK_RESULTS_DIR/results/ + + old_asv_dir=$previous_results/benchmarks/results/asv + if [ -d $old_asv_dir ]; then + cp -r $old_asv_dir $BENCHMARK_RESULTS_DIR/results + fi +} + + +# echos the name of the directory that $1 is linked to. Useful for +# getting the actual path of the results dir since that is often +# sym-linked to a unique (based on timestamp) results dir name. +function getNonLinkedFileName { + linkname=$1 + targetname=$(readlink -f $linkname) + if [[ "$targetname" != "" ]]; then + echo $targetname + else + echo $linkname + fi +} + +function waitForSlurmJobsToComplete { + ids=$* + jobs=$(python -c "print(\",\".join(\"$ids\".split()))") # make a comma-separated list + jobsInQueue=$(squeue --noheader --jobs=$jobs) + while [[ $jobsInQueue != "" ]]; do + sleep 2 + jobsInQueue=$(squeue --noheader --jobs=$jobs) + done +} + +# Clones repo from URL specified by $1 as name $2 in to directory +# $3. For example: +# "cloneRepo https://github.com/rapidsai/cugraph.git /my/repos cg" +# results in cugraph being cloned to /my/repos/cg. +# NOTE: This removes any existing cloned repos that match the +# destination. +function cloneRepo { + repo_url=$1 + repo_name=$2 + dest_dir=$3 + mkdir -p $dest_dir + pushd $dest_dir > /dev/null || exit + logger "Clone $repo_url in $dest_dir..." + if [ -d $repo_name ]; then + rm -rf $repo_name + if [ -d $repo_name ]; then + echo "ERROR: ${dest_dir}/$repo_name was not completely removed." + error 1 + fi + fi + git clone $repo_url + popd > /dev/null || exit +} + +# Only define this function if it has not already been defined in the +# current environment, which allows the project to override it from +# its functions.sh file that was previously source'd. +if [[ $(type -t activateCondaEnv) == "" ]]; then + function activateCondaEnv { + logger "Activating conda env ${CONDA_ENV}..." + eval "$(conda shell.bash hook)" + conda activate $CONDA_ENV + } +fi diff --git a/regression/get_datasets.py b/regression/get_datasets.py new file mode 100644 index 0000000000..bb2a9f23d3 --- /dev/null +++ b/regression/get_datasets.py @@ -0,0 +1,915 @@ +# SPDX-FileCopyrightText: Copyright (c) 2021-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. # noqa +# SPDX-License-Identifier: Apache-2.0 + +import os +import sys +import urllib.request +import urllib.parse +import ssl +import subprocess + + +# From: https://plato.asu.edu/bench.html +# Folder containg instances: +# - https://miplib2010.zib.de/miplib2010.php +# - https://www.netlib.org/lp/data/ +# - https://old.sztaki.hu/~meszaros/public_ftp/lptestset/ (and it's subfolders) +# - https://plato.asu.edu/ftp/lptestset/ (and it's subfolders) +# - https://miplib.zib.de/tag_benchmark.html +# - https://miplib.zib.de/tag_collection.html + +LPFeasibleMittelmannSet = [ + "L1_sixm250obs", + "Linf_520c", + "a2864", + "bdry2", + "cont1", + "cont11", + "datt256_lp", + "dlr1", + "ex10", + "fhnw-binschedule1", + "fome13", + "graph40-40", + "irish-electricity", + "neos", + "neos3", + "neos-3025225", + "neos-5052403-cygnet", + "neos-5251015", + "ns1687037", + "ns1688926", + "nug08-3rd", + "pds-100", + "physiciansched3-3", + "qap15", + "rail02", + "rail4284", + "rmine15", + "s82", + "s100", + "s250r10", + "savsched1", + "scpm1", + "shs1023", + "square41", + "stat96v2", + "stormG2_1000", + "stp3d", + "supportcase10", + "tpl-tub-ws1617", + "woodlands09", + "Dual2_5000", + "Primal2_1000", + "thk_48", + "thk_63", + "L1_sixm1000obs", + "L2CTA3D", + "degme", + "dlr2", + "set-cover-model", +] + +MiplibInstances = [ + "30n20b8.mps", + "cryptanalysiskb128n5obj14.mps", + "graph20-20-1rand.mps", + "n2seq36q.mps", + "neos-4338804-snowy.mps", + "neos-957323.mps", + "rail01.mps", + "splice1k1.mps", + "50v-10.mps", + "cryptanalysiskb128n5obj16.mps", + "graphdraw-domain.mps", + "n3div36.mps", + "neos-4387871-tavua.mps", + "neos-960392.mps", + "rail02.mps", + "square41.mps", + "academictimetablesmall.mps", + "csched007.mps", + "h80x6320d.mps", + "n5-3.mps", + "neos-4413714-turia.mps", + "net12.mps", + "rail507.mps", + "square47.mps", + "air05.mps", + "csched008.mps", + "highschool1-aigio.mps", + "neos-1122047.mps", + "neos-4532248-waihi.mps", + "netdiversion.mps", + "ran14x18-disj-8.mps", + "supportcase10.mps", + "app1-1.mps", + "cvs16r128-89.mps", + "hypothyroid-k1.mps", + "neos-1171448.mps", + "neos-4647030-tutaki.mps", + "nexp-150-20-8-5.mps", + "rd-rplusc-21.mps", + "supportcase12.mps", + "app1-2.mps", + "dano3_3.mps", + "ic97_potential.mps", + "neos-1171737.mps", + "neos-4722843-widden.mps", + "ns1116954.mps", + "reblock115.mps", + "supportcase18.mps", + "assign1-5-8.mps", + "dano3_5.mps", + "icir97_tension.mps", + "neos-1354092.mps", + "neos-4738912-atrato.mps", + "ns1208400.mps", + "rmatr100-p10.mps", + "supportcase19.mps", + "atlanta-ip.mps", + "decomp2.mps", + "irish-electricity.mps", + "neos-1445765.mps", + "neos-4763324-toguru.mps", + "ns1644855.mps", + "rmatr200-p5.mps", + "supportcase22.mps", + "b1c1s1.mps", + "drayage-100-23.mps", + "irp.mps", + "neos-1456979.mps", + "neos-4954672-berkel.mps", + "ns1760995.mps", + "rocI-4-11.mps", + "supportcase26.mps", + "bab2.mps", + "drayage-25-23.mps", + "istanbul-no-cutoff.mps", + "neos-1582420.mps", + "neos-5049753-cuanza.mps", + "ns1830653.mps", + "rocII-5-11.mps", + "supportcase33.mps", + "bab6.mps", + "dws008-01.mps", + "k1mushroom.mps", + "neos17.mps", + "neos-5052403-cygnet.mps", + "ns1952667.mps", + "rococoB10-011000.mps", + "supportcase40.mps", + "beasleyC3.mps", + "eil33-2.mps", + "lectsched-5-obj.mps", + "neos-2075418-temuka.mps", + "neos-5093327-huahum.mps", + "nu25-pr12.mps", + "rococoC10-001000.mps", + "supportcase42.mps", + "binkar10_1.mps", + "eilA101-2.mps", + "leo1.mps", + "neos-2657525-crna.mps", + "neos-5104907-jarama.mps", + "neos-5104907-jarama.mps", + "nursesched-medium-hint03.mps", + "roi2alpha3n4.mps", + "supportcase6.mps", + "blp-ar98.mps", + "enlight_hard.mps", + "leo2.mps", + "neos-2746589-doon.mps", + "neos-5107597-kakapo.mps", + "nursesched-sprint02.mps", + "roi5alpha10n8.mps", + "supportcase7.mps", + "blp-ic98.mps", + "ex10.mps", + "lotsize.mps", + "neos-2978193-inde.mps", + "neos-5114902-kasavu.mps", + "nw04.mps", + "roll3000.mps", + "swath1.mps", + "bnatt400.mps", + "ex9.mps", + "mad.mps", + "neos-2987310-joes.mps", + "neos-5188808-nattai.mps", + "opm2-z10-s4.mps", + "s100.mps", + "swath3.mps", + "bnatt500.mps", + "exp-1-500-5-5.mps", + "map10.mps", + "neos-3004026-krka.mps", + "neos-5195221-niemur.mps", + "p200x1188c.mps", + "s250r10.mps", + "tbfp-network.mps", + "bppc4-08.mps", + "fast0507.mps", + "map16715-04.mps", + "neos-3024952-loue.mps", + "neos5.mps", + "peg-solitaire-a3.mps", + "satellites2-40.mps", + "thor50dday.mps", + "brazil3.mps", + "fastxgemm-n2r6s0t2.mps", + "markshare2.mps", + "neos-3046615-murg.mps", + "neos-631710.mps", + "pg5_34.mps", + "satellites2-60-fs.mps", + "timtab1.mps", + "buildingenergy.mps", + "fhnw-binpack4-48.mps", + "markshare_4_0.mps", + "neos-3083819-nubu.mps", + "neos-662469.mps", + "pg.mps", + "savsched1.mps", + "tr12-30.mps", + "cbs-cta.mps", + "fhnw-binpack4-4.mps", + "mas74.mps", + "neos-3216931-puriri.mps", + "neos-787933.mps", + "physiciansched3-3.mps", + "sct2.mps", + "traininstance2.mps", + "chromaticindex1024-7.mps", + "fiball.mps", + "mas76.mps", + "neos-3381206-awhea.mps", + "neos-827175.mps", + "physiciansched6-2.mps", + "seymour1.mps", + "traininstance6.mps", + "chromaticindex512-7.mps", + "gen-ip002.mps", + "mc11.mps", + "neos-3402294-bobin.mps", + "neos-848589.mps", + "piperout-08.mps", + "seymour.mps", + "trento1.mps", + "cmflsp50-24-8-8.mps", + "gen-ip054.mps", + "mcsched.mps", + "neos-3402454-bohle.mps", + "neos859080.mps", + "piperout-27.mps", + "sing326.mps", + "triptim1.mps", + "CMS750_4.mps", + "germanrr.mps", + "mik-250-20-75-4.mps", + "neos-3555904-turama.mps", + "neos-860300.mps", + "pk1.mps", + "sing44.mps", + "uccase12.mps", + "co-100.mps", + "gfd-schedulen180f7d50m30k18.mps", + "milo-v12-6-r2-40-1.mps", + "neos-3627168-kasai.mps", + "neos-873061.mps", + "proteindesign121hz512p9.mps", + "snp-02-004-104.mps", + "uccase9.mps", + "cod105.mps", + "glass4.mps", + "momentum1.mps", + "neos-3656078-kumeu.mps", + "neos8.mps", + "proteindesign122trx11p8.mps", + "sorrell3.mps", + "uct-subprob.mps", + "comp07-2idx.mps", + "glass-sc.mps", + "mushroom-best.mps", + "neos-3754480-nidda.mps", + "neos-911970.mps", + "qap10.mps", + "sp150x300d.mps", + "unitcal_7.mps", + "comp21-2idx.mps", + "gmu-35-40.mps", + "mzzv11.mps", + "neos-3988577-wolgan.mps", + "neos-933966.mps", + "radiationm18-12-05.mps", + "sp97ar.mps", + "var-smallemery-m6j6.mps", + "cost266-UUE.mps", + "gmu-35-50.mps", + "mzzv42z.mps", + "neos-4300652-rahue.mps", + "neos-950242.mps", + "radiationm40-10-02.mps", + "sp98ar.mps", + "wachplan.mps", +] + +MittelmannInstances = { + "emps": "https://old.sztaki.hu/~meszaros/public_ftp/lptestset/emps.c", + "problems": { + "irish-electricity": [ + "https://plato.asu.edu/ftp/lptestset/irish-electricity.mps.bz2", + "mps", + ], + "physiciansched3-3": [ + "https://plato.asu.edu/ftp/lptestset/physiciansched3-3.mps.bz2", + "mps", + ], + "16_n14": [ + "https://plato.asu.edu/ftp/lptestset/network/16_n14.mps.bz2", + "mps", + ], + "Dual2_5000": [ + "https://plato.asu.edu/ftp/lptestset/Dual2_5000.mps.bz2", + "mps", + ], + "L1_six1000": [ + "https://plato.asu.edu/ftp/lptestset/L1_sixm1000obs.bz2", + "netlib", + ], + "L1_sixm": ["", "mps"], + "L1_sixm1000obs": [ + "https://plato.asu.edu/ftp/lptestset/L1_sixm1000obs.bz2", + "netlib", + ], + "L1_sixm250": ["", "netlib"], + "L1_sixm250obs": [ + "https://plato.asu.edu/ftp/lptestset/L1_sixm250obs.bz2", + "netlib", + ], + "L2CTA3D": [ + "https://plato.asu.edu/ftp/lptestset/L2CTA3D.mps.bz2", + "mps", + ], + "Linf_520c": [ + "https://plato.asu.edu/ftp/lptestset/Linf_520c.bz2", + "netlib", + ], + "Primal2_1000": [ + "https://plato.asu.edu/ftp/lptestset/Primal2_1000.mps.bz2", + "mps", + ], + "a2864": ["https://plato.asu.edu/ftp/lptestset/a2864.mps.bz2", "mps"], + "bdry2": ["https://plato.asu.edu/ftp/lptestset/bdry2.bz2", "netlib"], + "braun": ["", "mps"], + "cont1": [ + "https://plato.asu.edu/ftp/lptestset/misc/cont1.bz2", + "netlib", + ], + "cont11": [ + "https://plato.asu.edu/ftp/lptestset/misc/cont11.bz2", + "netlib", + ], + "datt256": [ + "https://plato.asu.edu/ftp/lptestset/datt256_lp.mps.bz2", + "mps", + ], + "datt256_lp": [ + "https://plato.asu.edu/ftp/lptestset/datt256_lp.mps.bz2", + "mps", + ], + "degme": [ + "https://old.sztaki.hu/~meszaros/public_ftp/lptestset/New/degme.gz", + "netlib", + ], + "dlr1": ["https://plato.asu.edu/ftp/lptestset/dlr1.mps.bz2", "mps"], + "dlr2": ["https://plato.asu.edu/ftp/lptestset/dlr2.mps.bz2", "mps"], + "energy1": ["", "mps"], # Kept secret by Mittlemman + "energy2": ["", "mps"], + "ex10": ["https://plato.asu.edu/ftp/lptestset/ex10.mps.bz2", "mps"], + "fhnw-binschedule1": [ + "https://plato.asu.edu/ftp/lptestset/fhnw-binschedule1.mps.bz2", + "mps", + ], + "fome13": [ + "https://plato.asu.edu/ftp/lptestset/fome/fome13.bz2", + "netlib", + ], + "gamora": ["", "mps"], # Kept secret by Mittlemman + "goto14_256_1": ["", "mps"], + "goto14_256_2": ["", "mps"], + "goto14_256_3": ["", "mps"], + "goto14_256_4": ["", "mps"], + "goto14_256_5": ["", "mps"], + "goto16_64_1": ["", "mps"], + "goto16_64_2": ["", "mps"], + "goto16_64_3": ["", "mps"], + "goto16_64_4": ["", "mps"], + "goto16_64_5": ["", "mps"], + "goto32_512_1": ["", "mps"], + "goto32_512_2": ["", "mps"], + "goto32_512_3": ["", "mps"], + "goto32_512_4": ["", "mps"], + "goto32_512_5": ["", "mps"], + "graph40-40": [ + "https://plato.asu.edu/ftp/lptestset/graph40-40.mps.bz2", + "mps", + ], + "graph40-40_lp": [ + "https://plato.asu.edu/ftp/lptestset/graph40-40.mps.bz2", + "mps", + ], + "groot": ["", "mps"], # Kept secret by Mittlemman + "heimdall": ["", "mps"], # Kept secret by Mittlemman + "hulk": ["", "mps"], # Kept secret by Mittlemman + "i_n13": [ + "https://plato.asu.edu/ftp/lptestset/network/i_n13.mps.bz2", + "mps", + ], + "irish-e": ["", "mps"], + "karted": [ + "https://old.sztaki.hu/~meszaros/public_ftp/lptestset/New/karted.gz", + "netlib", + ], + "lo10": [ + "https://plato.asu.edu/ftp/lptestset/network/lo10.mps.bz2", + "mps", + ], + "loki": ["", "mps"], # Kept secret by Mittlemman + "long15": [ + "https://plato.asu.edu/ftp/lptestset/network/long15.mps.bz2", + "mps", + ], + "nebula": ["", "mps"], # Kept secret by Mittlemman + "neos": [ + "https://plato.asu.edu/ftp/lptestset/misc/neos.bz2", + "netlib", + ], + "neos-3025225": [ + "https://plato.asu.edu/ftp/lptestset/neos-3025225.mps.bz2", + "mps", + ], + "neos-3025225_lp": [ + "https://plato.asu.edu/ftp/lptestset/neos-3025225.mps.bz2", + "mps", + ], + "neos-5251015": [ + "https://plato.asu.edu/ftp/lptestset/neos-5251015.mps.bz2", + "mps", + ], + "neos-5251015_lp": [ + "https://plato.asu.edu/ftp/lptestset/neos-5251015.mps.bz2", + "mps", + ], + "neos3": [ + "https://plato.asu.edu/ftp/lptestset/misc/neos3.bz2", + "netlib", + ], + "neos-5052403-cygnet": [ + "https://plato.asu.edu/ftp/lptestset/neos-5052403-cygnet.mps.bz2", + "mps", + ], + "neos5251015_lp": [ + "https://plato.asu.edu/ftp/lptestset/neos-5251015.mps.bz2", + "mps", + ], + "neos5251915": [ + "https://plato.asu.edu/ftp/lptestset/neos-5251015.mps.bz2", + "mps", + ], + "netlarge1": [ + "https://plato.asu.edu/ftp/lptestset/network/netlarge1.mps.bz2", + "mps", + ], + "netlarge2": [ + "https://plato.asu.edu/ftp/lptestset/network/netlarge2.mps.bz2", + "mps", + ], + "netlarge3": [ + "https://plato.asu.edu/ftp/lptestset/network/netlarge3.mps.bz2", + "mps", + ], + "netlarge6": [ + "https://plato.asu.edu/ftp/lptestset/network/netlarge6.mps.bz2", + "mps", + ], + "ns1687037": [ + "https://plato.asu.edu/ftp/lptestset/misc/ns1687037.bz2", + "netlib", + ], + "ns1688926": [ + "https://plato.asu.edu/ftp/lptestset/misc/ns1688926.bz2", + "netlib", + ], + "nug08-3rd": [ + "https://plato.asu.edu/ftp/lptestset/nug/nug08-3rd.bz2", + "netlib", + ], + "pds-100": [ + "https://plato.asu.edu/ftp/lptestset/pds/pds-100.bz2", + "netlib", + ], + "psched3-3": ["", "mps"], + "qap15": ["https://plato.asu.edu/ftp/lptestset/qap15.mps.bz2", "mps"], + "rail02": ["https://miplib2010.zib.de/download/rail02.mps.gz", "mps"], + "rail4284": [ + "https://plato.asu.edu/ftp/lptestset/rail/rail4284.bz2", + "netlib", + ], + "rmine15": [ + "https://plato.asu.edu/ftp/lptestset/rmine15.mps.bz2", + "mps", + ], + "s100": ["https://plato.asu.edu/ftp/lptestset/s100.mps.bz2", "mps"], + "s250r10": [ + "https://plato.asu.edu/ftp/lptestset/s250r10.mps.bz2", + "mps", + ], + "s82": ["https://plato.asu.edu/ftp/lptestset/s82.mps.bz2", "mps"], + "savsched1": [ + "https://plato.asu.edu/ftp/lptestset/savsched1.mps.bz2", + "mps", + ], + "scpm1": ["https://plato.asu.edu/ftp/lptestset/scpm1.mps.bz2", "mps"], + "set-cover-model": [ + "https://plato.asu.edu/ftp/lptestset/set-cover-model.mps.bz2", + "mps", + ], + "shs1023": [ + "https://miplib2010.zib.de/download/shs1023.mps.gz", + "mps", + ], + "square15": [ + "https://plato.asu.edu/ftp/lptestset/network/square15.mps.bz2", + "mps", + ], + "square41": [ + "https://plato.asu.edu/ftp/lptestset/square41.mps.bz2", + "mps", + ], + "stat96v2": [ + "https://old.sztaki.hu/~meszaros/public_ftp/lptestset/misc/stat96v2.gz", + "netlib", + ], + "stormG2_1000": [ + "https://plato.asu.edu/ftp/lptestset/misc/stormG2_1000.bz2", + "netlib", + ], + "storm_1000": ["", "mps"], + "stp3d": [ + "https://miplib.zib.de/WebData/instances/stp3d.mps.gz", + "mps", + ], + "supportcase10": [ + "https://plato.asu.edu/ftp/lptestset/supportcase10.mps.bz2", + "mps", + ], + "support19": [ + "https://plato.asu.edu/ftp/lptestset/supportcase19.mps.bz2", + "mps", + ], + "supportcase19": [ + "https://plato.asu.edu/ftp/lptestset/supportcase19.mps.bz2", + "mps", + ], + "test03": ["", "mps"], # Kept secret by Mittlemman + "test13": ["", "mps"], # Kept secret by Mittlemman + "test23": ["", "mps"], # Kept secret by Mittlemman + "test33": ["", "mps"], # Kept secret by Mittlemman + "test43": ["", "mps"], # Kept secret by Mittlemman + "test53": ["", "mps"], # Kept secret by Mittlemman + "test63": ["", "mps"], # Kept secret by Mittlemman + "test83": ["", "mps"], # Kept secret by Mittlemman + "test93": ["", "mps"], # Kept secret by Mittlemman + "mars": ["", "mps"], # Kept secret by Mittlemman + "thk_48": [ + "https://plato.asu.edu/ftp/lptestset/thk_48.mps.bz2", + "mps", + ], + "thk_63": [ + "https://plato.asu.edu/ftp/lptestset/thk_63.mps.bz2", + "mps", + ], + "thor": ["", "mps"], # Kept secret by Mittlemman + "tpl-tub-ws": ["", "mps"], + "tpl-tub-ws1617": [ + "https://plato.asu.edu/ftp/lptestset/tpl-tub-ws1617.mps.bz2", + "mps", + ], + "wide15": [ + "https://plato.asu.edu/ftp/lptestset/network/wide15.mps.bz2", + "mps", + ], + "woodlands09": [ + "https://plato.asu.edu/ftp/lptestset/woodlands09.mps.bz2", + "mps", + ], + }, + "benchmarks": { + "simplex": [ + "L1_sixm", + "L1_sixm250obs", + "Linf_520c", + "a2864", + "bdry2", + "braun", + "cont1", + "cont11", + "datt256", + "dlr1", + "energy1", + "energy2", + "ex10", + "fhnw-binschedule1", + "fome13", + "gamora", + "graph40-40", + "groot", + "heimdall", + "hulk", + "irish-e", + "loki", + "nebula", + "neos", + "neos-3025225_lp", + "neos-5251015_lp", + "neos3", + "neos3025225", + "neos5052403", + "neos5251015_lp", + "ns1687037", + "ns1688926", + "nug08-3rd", + "pds-100", + "psched3-3", + "qap15", + "rail02", + "rail4284", + "rmine15", + "s100", + "s250r10", + "s82", + "savsched1", + "scpm1", + "shs1023", + "square41", + "stat96v2", + "stormG2_1000", + "storm_1000", + "stp3d", + "support10", + "test03", + "test13", + "test23", + "test33", + "test43", + "test53", + "thor", + "tpl-tub-ws", + "tpl-tub-ws16", + "woodlands09", + ], + "barrier": [ + "Dual2_5000", + "L1_six1000", + "L1_sixm1000obs", + "L1_sixm250", + "L1_sixm250obs", + "L2CTA3D", + "Linf_520c", + "Primal2_1000", + "a2864", + "bdry2", + "cont1", + "cont11", + "datt256", + "degme", + "dlr1", + "dlr2", + "ex10", + "fhnw-binschedule1", + "fome13", + "graph40-40", + "irish-e", + "karted", + "neos", + "neos-3025225_lp", + "neos-5251015_lp", + "neos3", + "neos3025225", + "neos5052403", + "neos5251915", + "ns1687037", + "ns1688926", + "nug08-3rd", + "pds-100", + "psched3-3", + "qap15", + "rail02", + "rail4284", + "rmine15", + "s100", + "s250r10", + "s82", + "savsched1", + "scpm1", + "set-cover-model", + "shs1023", + "square41", + "stat96v2", + "stormG2_1000", + "storm_1000", + "stp3d", + "support10", + "support19", + "supportcase19", + "thk_63", + "tpl-tub-ws", + "tpl-tub-ws16", + "woodlands09", + ], + "large": [ + "16_n14", + "goto14_256_1", + "goto14_256_2", + "goto14_256_3", + "goto14_256_4", + "goto14_256_5", + "goto16_64_1", + "goto16_64_2", + "goto16_64_3", + "goto16_64_4", + "goto16_64_5", + "goto32_512_1", + "goto32_512_2", + "goto32_512_3", + "goto32_512_4", + "goto32_512_5", + "i_n13", + "lo10", + "long15", + "netlarge1", + "netlarge2", + "netlarge3", + "netlarge6", + "square15", + "wide15", + ], + # <=100s in bench: https://plato.asu.edu/ftp/lpbar.html + "L0": [ + "ex10", + "datt256", + "graph40-40", + "neos5251915", + "nug08-3rd", + "qap15", + "savsched1", + "scpm1", + "a2864", + "support10", + "rmine15", + "fome13", + "L2CTA3D", + "neos5052403", + "karted", + "stp3d", + "woodlands09", + "rail4284", + "L1_sixm250", + "tpl-tub-ws", + ], + # >100 <1000 + "L1": [ + "s250r10", + "pds-100", + "set-cover-model", + "neos3025225", + "rail02", + "square41", + "degme", + "Linf_520c", + "cont1", + "neos", + "stat96v2", + "support19", + "shs1023", + "storm_1000", + ], + # >1000 + "L2": [ + "thk_63", + "Primal2_1000", + "L1_six1000", + "Dual2_5000", + "s100", + "fhnw-binschedule1", + "cont11", + "psched3-3", + ], + # t -> >15000 + "L3": [ + "dlr2", + "bdry2", + "dlr1", + "irish-e", + "ns1687037", + "ns1688926", + "s82", + ], + }, +} + + +def download(url, dst): + if os.path.exists(dst): + return + print(f"Downloading {url} into {dst}...") + # Bypass SSL verification for plato.asu.edu URLs + if "plato.asu.edu" in url: + context = ssl.create_default_context() + context.check_hostname = False + context.verify_mode = ssl.CERT_NONE + response = urllib.request.urlopen(url, context=context) + else: + response = urllib.request.urlopen(url) + data = response.read() + with open(dst, "wb") as fp: + fp.write(data) + + +def extract(file, dir, type): + basefile = os.path.basename(file) + outfile = "" + unzippedfile = "" + if basefile.endswith(".bz2"): + outfile = basefile.replace(".bz2", ".mps") + unzippedfile = basefile.replace(".bz2", "") + subprocess.run(f"cd {dir} && bzip2 -d {basefile}", shell=True) + elif basefile.endswith(".gz"): + outfile = basefile.replace(".gz", ".mps") + unzippedfile = basefile.replace(".gz", "") + subprocess.run( + f"cd {dir} && gunzip -c {basefile} > {unzippedfile}", shell=True + ) + subprocess.run(f"cd {dir} && rm -rf {basefile}", shell=True) + else: + raise Exception(f"Unknown file extension found for extraction {file}") + # download emps and compile + # Disable emps for now + if type == "netlib": + url = MittelmannInstances["emps"] + file = os.path.join(dir, "emps.c") + download(url, file) + subprocess.run( + f"cd {dir} && gcc -Wno-implicit-int emps.c -o emps", shell=True + ) + # determine output file and run emps + subprocess.run( + f"cd {dir} && ./emps {unzippedfile} > {outfile}", shell=True + ) + subprocess.run(f"cd {dir} && rm -rf {unzippedfile}", shell=True) + # cleanup emps and emps.c + subprocess.run(f"rm -rf {dir}/emps*", shell=True) + + +def download_lp_dataset(name, dir): + if name not in MittelmannInstances["problems"]: + raise Exception(f"Unknown dataset {name} passed") + if os.path.exists(dir): + if os.path.exists(os.path.join(dir, f"{name}.mps")): + print( + f"Dir for dataset {name} exists and contains {name}.mps. Skipping..." + ) + return + url, type = MittelmannInstances["problems"][name] + if url == "": + print(f"Dataset {name} doesn't have a URL. Skipping...") + return + file = os.path.join(dir, os.path.basename(url)) + download(url, file) + extract(file, dir, type) + + +def download_mip_dataset(name, dir): + base_url = "https://miplib.zib.de/WebData/instances" + url = f"{base_url}/{name}.gz" + outfile = f"{dir}/{name}.gz" + if os.path.exists(dir): + if os.path.exists(os.path.join(dir, f"{name}")): + print( + f"Dir for dataset {name} exists and contains {name}.mps. Skipping..." + ) + return + download(url, outfile) + extract(outfile, dir, "") + + +datasets_path = sys.argv[1] +dataset_type = sys.argv[2] + +if dataset_type == "lp": + for name in LPFeasibleMittelmannSet: + download_lp_dataset(name, datasets_path) +elif dataset_type == "mip": + for name in MiplibInstances: + download_mip_dataset(name, datasets_path) diff --git a/regression/lp_config.json b/regression/lp_config.json new file mode 100644 index 0000000000..e2f8a9e936 --- /dev/null +++ b/regression/lp_config.json @@ -0,0 +1,13 @@ +{ + "details": "LP test", + "metrics": { + "primal_objective_value": { + "threshold": 1, + "unit": "primal_objective_value" + }, + "solver_time": { + "threshold": 1, + "unit": "seconds" + } + } +} diff --git a/regression/lp_regression_test.sh b/regression/lp_regression_test.sh new file mode 100644 index 0000000000..23b28cd1f5 --- /dev/null +++ b/regression/lp_regression_test.sh @@ -0,0 +1,29 @@ +#!/bin/bash +# SPDX-FileCopyrightText: Copyright (c) 2021-2026, NVIDIA CORPORATION. All rights reserved. +# SPDX-License-Identifier: Apache-2.0 + +# Abort script on first error +set -e + +# Must ensure PROJECT_DIR is exported first then load rapids-mg-tools env +export PROJECT_DIR=${PROJECT_DIR:-$(cd "$(dirname ${BASH_SOURCE[0]})" && pwd)} +source ${PROJECT_DIR}/config.sh +source ${PROJECT_DIR}/functions.sh + +################################################################################ + +# Extract the build meta-data from either the conda environment or the +# cugraph source dir and write out a file which can be read by other +# scripts. If the cugraph conda packages are present, those take +# precedence, otherwise meta-data will be extracted from the sources. + +GIT_COMMIT=$( /dev/null); then + STATUS='PASSED' + STATUS_IMG='https://img.icons8.com/bubbles/100/000000/approval.png' + fi + +fi + +# Generate a one-line summary based on existance of certain reports, etc. +if [[ "$ALL_REPORTS" == "" ]]; then + ONE_LINE_SUMMARY="*Build failed*" +elif [[ "$STATUS" == "FAILED" ]]; then + if (grep -w FAILED $BENCHMARK_REPORT > /dev/null); then + ONE_LINE_SUMMARY="*One or more benchmarks failed*" + fi + if (grep -w FAILED $TEST_REPORT > /dev/null); then + ONE_LINE_SUMMARY="*One or more tests failed*" + fi + if (grep -w FAILED $TEST_REPORT > /dev/null) && (grep -w FAILED $BENCHMARK_REPORT > /dev/null); then + ONE_LINE_SUMMARY="*One or more tests and benchmarks failed*" + fi +else + ONE_LINE_SUMMARY="Build succeeded, all tests and benchmarks passed" +fi + +RESULTS_DIR_NAME=$(basename "$(getNonLinkedFileName $RESULTS_DIR)") + +# Upload everything +logger "Uploading all files in $RESULTS_DIR ..." +logger "Uploading all files in $RESULTS_DIR_NAME ..." +aws s3 cp --follow-symlinks --acl public-read --recursive ${RESULTS_DIR} ${S3_FILE_PREFIX}/${RESULTS_DIR_NAME} +logger "done uploading all files in $RESULTS_DIR" + +# Set vars used in the report +PROJECT_VERSION_STRING="" +PROJECT_VERSION="" +PROJECT_BUILD="" +PROJECT_CHANNEL="" +PROJECT_REPO_URL="" +PROJECT_REPO_BRANCH="" +if [ -f $METADATA_FILE ]; then + source $METADATA_FILE +fi +# Assume if PROJECT_BUILD is set then a conda version string should be +# created, else a git version string. +if [[ "$PROJECT_BUILD" != "" ]]; then + PROJECT_VERSION_STRING=" cuOpt ver.: $PROJECT_VERSION + build: $PROJECT_BUILD + channel: $PROJECT_CHANNEL" +else + PROJECT_VERSION_STRING=" cuOpt ver.: $PROJECT_VERSION + repo: $PROJECT_REPO_URL + branch: $PROJECT_REPO_BRANCH" +fi + +export STATUS +export STATUS_IMG +export PROJECT_VERSION_STRING +HUMAN_READABLE_DATE="$(date '+`%D`, `%H:%M` (PT)')" +export HUMAN_READABLE_DATE +# These files should be created by create-html-reports.sh +export REPORT_URL="${S3_URL_PREFIX}/${RESULTS_DIR_NAME}/report.html" +export ASV_URL="${S3_URL_PREFIX}/${RESULTS_DIR_NAME}/benchmarks/asv/html/index.html" +export LOGS_URL="${S3_URL_PREFIX}/${RESULTS_DIR_NAME}/index.html" +# export SPREADSHEET_URL=$SPREADSHEET_URL +export ONE_LINE_SUMMARY + +echo +echo "REPORT_URL: ${REPORT_URL}" +# echo "SPREADSHEET_URL: ${SPREADSHEET_URL}" + +if hasArg --skip-sending-report; then + logger "Skipping sending Slack report." +else + echo "$(envsubst < ${PROJECT_DIR}/slack_msg.json)" + curl -X POST \ + -H 'Content-type: application/json' \ + --data "$(envsubst < ${PROJECT_DIR}/slack_msg.json)" \ + ${WEBHOOK_URL} +fi diff --git a/regression/slack_msg.json b/regression/slack_msg.json new file mode 100644 index 0000000000..a73e659b2b --- /dev/null +++ b/regression/slack_msg.json @@ -0,0 +1,68 @@ +{ + "channel": "cuopt-regression-testing", + "username": "cuOpt Messaging", + "icon_emoji": ":robot_face:", + "blocks": [ + { + "type": "section", + "text": { + "type": "mrkdwn", + "text": "${ONE_LINE_SUMMARY}" + } + }, + { + "type": "divider" + }, + { + "type": "section", + "text": { + "type": "mrkdwn", + "text": "<${REPORT_URL}|*Results Report*>\nBuild status and test results." + }, + "accessory": { + "type": "button", + "url": "${REPORT_URL}", + "text": { + "type": "plain_text", + "emoji": true, + "text": "View" + }, + "value": "click_me_123" + } + }, + { + "type": "section", + "text": { + "type": "mrkdwn", + "text": "<${ASV_URL}|*ASV Dashboard*>\nBenchmark results." + }, + "accessory": { + "type": "button", + "url": "${ASV_URL}", + "text": { + "type": "plain_text", + "emoji": true, + "text": "View" + }, + "value": "click_me_123" + } + }, + { + "type": "section", + "text": { + "type": "mrkdwn", + "text": "<${LOGS_URL}|*Logs*>\nAll available logs." + }, + "accessory": { + "type": "button", + "url": "${LOGS_URL}", + "text": { + "type": "plain_text", + "emoji": true, + "text": "View" + }, + "value": "click_me_123" + } + } + ] +} diff --git a/regression/test-container.sh b/regression/test-container.sh new file mode 100644 index 0000000000..4aeacb5678 --- /dev/null +++ b/regression/test-container.sh @@ -0,0 +1,26 @@ +#!/bin/bash +# SPDX-FileCopyrightText: Copyright (c) 2021-2026, NVIDIA CORPORATION. All rights reserved. +# SPDX-License-Identifier: Apache-2.0 + +# Creates a conda environment to be used for cuopt benchmarking. + +# Abort script on first error +set -e + +# Must ensure PROJECT_DIR is exported first then load rapids-mg-tools env +export PROJECT_DIR=${PROJECT_DIR:-$(cd "$(dirname ${BASH_SOURCE[0]})" && pwd)} +source ${PROJECT_DIR}/config.sh + +################################################################################ + +# Test +logger "Testing container image $IMAGE" +python -c "import cuopt; print(cuopt)" + +trap '${SCRIPTS_DIR}/write-meta-data.sh' EXIT + +# Other scripts look for this to be the last line to determine if this +# script completed successfully. This is only possible because of the +# "set -e" above. +echo "done." +logger "done." diff --git a/regression/update_asv_database.py b/regression/update_asv_database.py new file mode 100644 index 0000000000..bef64bd313 --- /dev/null +++ b/regression/update_asv_database.py @@ -0,0 +1,157 @@ +# SPDX-FileCopyrightText: Copyright (c) 2021-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. +# SPDX-License-Identifier: Apache-2.0 + +from pathlib import Path +import platform +import psutil +from asvdb import BenchmarkInfo, BenchmarkResult, ASVDb +import json +import pandas as pd + + +def update_asv_db( + commitHash=None, + commitTime=None, + branch=None, + repo_url=None, + results_dir=None, + machine_name=None, + gpu_type=None, + configs=None, +): + """ + Read the benchmark_result* files in results_dir/benchmarks and + update an existing asv benchmark database or create one if one + does not exist in results_dir/benchmarks/asv. If no + benchmark_result* files are present, return without updating or + creating. + """ + + # commitHash = commitHash + str(int(time.time())) + benchmark_dir_path = Path(results_dir) / "benchmarks" / "results" / "csvs" + asv_dir_path = Path(results_dir) / "benchmarks" / "results" / "asv" + + # List all benchmark_result files + benchmark_result_list = benchmark_dir_path.glob("*.csv") + + bResultList = [] + # Skip these columns from benchmarking + skip_columns = ["date_time", "git_commit"] + + # Create result objects for each benchmark result and store it in a list + for file_name in benchmark_result_list: + # skip if it's regression file + if "regressions.csv" in str(file_name): + continue + with open(file_name, "r") as openfile: + data = pd.read_csv(openfile, index_col=0).iloc[-1] + test_name = str(file_name).split("/")[-1].split(".")[-2] + config_file = None + if test_name.startswith("lp"): + config_file = configs + "/" + "lp_config.json" + elif test_name.startswith("mip"): + config_file = configs + "/" + "mip_config.json" + else: + config_file = configs + "/" + test_name + "_config.json" + metrics = {} + with open(config_file, "r") as fp: + metrics = json.load(fp)["metrics"] + for col_name in data.index: + if col_name not in skip_columns: + bResult = BenchmarkResult( + funcName=test_name + "." + col_name, + result=data[col_name].item(), + unit="percentage" + if "bks" in col_name + else metrics[col_name]["unit"], + ) + bResultList.append(bResult) + + if len(bResultList) == 0: + print( + "Could not find files matching 'csv' in " + f"{benchmark_dir_path}, not creating/updating ASV database " + f"in {asv_dir_path}." + ) + return + + uname = platform.uname() + # Maybe also write those metadata to metadata.sh ? + osType = "%s %s" % (uname.system, uname.release) + # Remove unnecessary osType detail + osType = ".".join(osType.split("-")[0].split(".", 2)[:2]) + pythonVer = platform.python_version() + # Remove unnecessary python version detail + pythonVer = ".".join(pythonVer.split(".", 2)[:2]) + bInfo_dict = { + "machineName": machine_name, + # cudaVer : "10.0", + "osType": osType, + "pythonVer": pythonVer, + "commitHash": commitHash, + "branch": branch, + # commit time needs to be in milliseconds + "commitTime": commitTime * 1000, + "gpuType": gpu_type, + "cpuType": uname.processor, + "arch": uname.machine, + "ram": "%d" % psutil.virtual_memory().total, + } + bInfo = BenchmarkInfo(**bInfo_dict) + + # extract the branch name + branch = bInfo_dict["branch"] + + db = ASVDb(dbDir=str(asv_dir_path), repo=repo_url, branches=[branch]) + + for res in bResultList: + db.addResult(bInfo, res) + + +if __name__ == "__main__": + import argparse + + ap = argparse.ArgumentParser() + ap.add_argument( + "--commitHash", type=str, required=True, help="project version" + ) + ap.add_argument( + "--commitTime", type=str, required=True, help="project version date" + ) + ap.add_argument( + "--repo-url", type=str, required=True, help="project repo url" + ) + ap.add_argument("--branch", type=str, required=True, help="project branch") + ap.add_argument( + "--results-dir", + type=str, + required=True, + help="directory to store the results in json files", + ) + ap.add_argument( + "--machine-name", type=str, required=True, help="Slurm cluster name" + ) + ap.add_argument( + "--gpu-type", + type=str, + required=True, + help="the official product name of the GPU", + ) + ap.add_argument( + "--configs", + type=str, + required=True, + help="the config file for all the tests", + ) + args = ap.parse_args() + + update_asv_db( + commitHash=args.commitHash, + commitTime=int(args.commitTime), + branch=args.branch, + repo_url=args.repo_url, + results_dir=args.results_dir, + machine_name=args.machine_name, + gpu_type=args.gpu_type, + configs=args.configs, + ) diff --git a/regression/write-meta-data.sh b/regression/write-meta-data.sh new file mode 100755 index 0000000000..020631469d --- /dev/null +++ b/regression/write-meta-data.sh @@ -0,0 +1,26 @@ +#!/bin/bash +# SPDX-FileCopyrightText: Copyright (c) 2021-2026, NVIDIA CORPORATION. All rights reserved. +# SPDX-License-Identifier: Apache-2.0 + +# Abort script on first error +set -e + +# Must ensure PROJECT_DIR is exported first then load rapids-mg-tools env +export PROJECT_DIR=${PROJECT_DIR:-$(cd "$(dirname ${BASH_SOURCE[0]})" && pwd)} +source ${PROJECT_DIR}/config.sh +source ${PROJECT_DIR}/functions.sh + +PROJECT_VERSION=$(> $METADATA_FILE +echo "PROJECT_VERSION=\"$PROJECT_VERSION\"" >> $METADATA_FILE +echo "PROJECT_BUILD=\"$PROJECT_BUILD\"" >> $METADATA_FILE +echo "PROJECT_CHANNEL=\"$PROJECT_CHANNEL\"" >> $METADATA_FILE +echo "PROJECT_REPO_URL=\"$PROJECT_REPO_URL\"" >> $METADATA_FILE +echo "PROJECT_REPO_BRANCH=\"$PROJECT_REPO_BRANCH\"" >> $METADATA_FILE +echo "PROJECT_REPO_TIME=\"$PROJECT_REPO_TIME\"" >> $METADATA_FILE From 701db7ed1890cace3173d816ab2f16bc6bc245b3 Mon Sep 17 00:00:00 2001 From: Chris Maes Date: Wed, 4 Feb 2026 20:12:13 -0800 Subject: [PATCH 2/3] Add cuts to the MIP solver (#814) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit This PR adds cuts to the MIP solver. This includes the following: 1. Add constraints in the form C*x <= d to an LP that has already been solved to optimality (and has basis information). - The constraints must be violated at the current relaxation solution x^star. That is, C*x^star > d. - The constraint matrix, rhs, basis, and basis factorization, are all updated to include the additional constraints. - Dual simplex is started in phase 2 from a dual feasible solution. 2. Remove constraints from an LP that has already been solved to optimality. - The constraints must have slacks in the basis - The basis is refactored from scratch 3. Add cut pass loop after solving the root relaxation 4. Add a cut pool to store cuts and select cuts - We currently score cuts based on distance and orthogonality. 6. Add Mixed Integer Gomory Cuts - These are computed via a MIR cut on a row of the simplex tableau 7. Add Mixed Integer Rounding (MIR) Cuts - These are constructed by aggregating rows of the constraint matrix. 8. Add Strong Chvatal-Gomory Cuts - These are constructed from a row of the tableau matrix and from rows of the constraint matrix. 9. Fixes to Handling of Steepest Edge Norms in Dual Simplex - Ensure that all basic variables have a positive steepest edge norms 10. Reduced Costs Fixing at Root Node - These are applied after each cut pass and after strong branching, if a heuristic solution is available. 12. Fix issues in Crossover when solving the dual problem - We were not correctly populating slack variables when solving the dual. This issue appeared on graph20-80-1rand 14. Fix issue in Crossover when basis became rank-deficient in dual push 15. Fix issues across the code with handling and propagating concurrent halt. 16. New solver options: mip-cut-passes, mip-mixed-integer-gomory-cuts, mip-mir-cuts, mip-strong-chvatal-gomory-cuts, mip-knapsack-cuts, mip-cut-change-threshold, mip-cut-min-orthogonality. Closes #698, #205 Results from a GH200 A/B test with 64 threads and a 300 second time limit. Further runs needed with larger time limit. Further work is needed to get the full benefit of cuts. A: cuts PR with --mip-cut-passes=10 B: cuts PR with --mip-cut-passes=0 Geomean MIP GAP A / (B = baseline): 0.97 Geomean Time to Optimal A/B: 0.96 A optimal 45 B optimal 37 A problems with feasible solutions 225 B problems with feasible solutions 224 A wins drayage-100-23 1.14 8.61 gfd-schedulen180f7d50m30k18 87.85 300.0 neos-827175 12.28 25.86 neos-1171448 28.85 98.2 n5-3 180.62 300.0 neos859080 0.78 1.38 seymour1 38.78 300.0 neos-860300 71.98 90.91 neos-3083819-nubu 30.48 300.0 neos-933966 248.26 300.0 neos-957323 46.36 300.0 neos-960392 50.12 115.68 netdiversion 121.59 300.0 ns1208400 162.15 300.0 nw04 39.89 45.98 piperout-27 8.22 14.09 supportcase7 30.48 300.0 supportcase6 70.22 300.0 uccase12 26.95 41.87 A wins 19 A losses app1-1 2.3 0.74 cbs-cta 2.21 1.83 irp 17.59 12.94 istanbul-no-cutoff 18.26 13.56 mas76 300.0 20.98 neos-1122047 8.3 6.46 neos-1445765 1.86 1.52 neos-1582420 134.3 93.44 neos-3004026-krka 6.56 1.31 neos8 0.88 0.68 ns1952667 99.84 64.92 piperout-08 10.75 8.23 pk1 224.32 91.84 qap10 134.76 62.82 swath1 60.04 50.3 trento1 300.0 172.28 triptim1 246.65 130.92 A losses 17 Authors: - Chris Maes (https://github.com/chris-maes) Approvers: - Ramakrishnap (https://github.com/rgsl888prabhu) - Akif ÇÖRDÜK (https://github.com/akifcorduk) - Alice Boucher (https://github.com/aliceb-nv) URL: https://github.com/NVIDIA/cuopt/pull/814 --- .../cuopt/linear_programming/constants.h | 10 +- .../mip/solver_settings.hpp | 18 +- cpp/src/dual_simplex/CMakeLists.txt | 1 + cpp/src/dual_simplex/barrier.cu | 10 +- cpp/src/dual_simplex/basis_solves.cpp | 4 +- cpp/src/dual_simplex/basis_updates.cpp | 255 +- cpp/src/dual_simplex/basis_updates.hpp | 2 + .../bound_flipping_ratio_test.cpp | 17 +- .../bound_flipping_ratio_test.hpp | 8 +- cpp/src/dual_simplex/bounds_strengthening.cpp | 36 +- cpp/src/dual_simplex/bounds_strengthening.hpp | 13 +- cpp/src/dual_simplex/branch_and_bound.cpp | 804 ++++- cpp/src/dual_simplex/branch_and_bound.hpp | 34 +- cpp/src/dual_simplex/crossover.cpp | 94 +- cpp/src/dual_simplex/cuts.cpp | 2822 +++++++++++++++++ cpp/src/dual_simplex/cuts.hpp | 479 +++ cpp/src/dual_simplex/dense_matrix.hpp | 4 +- cpp/src/dual_simplex/folding.cpp | 71 +- cpp/src/dual_simplex/mip_node.hpp | 9 +- cpp/src/dual_simplex/phase2.cpp | 211 +- cpp/src/dual_simplex/presolve.cpp | 9 +- cpp/src/dual_simplex/primal.cpp | 10 +- cpp/src/dual_simplex/primal.hpp | 6 +- cpp/src/dual_simplex/pseudo_costs.cpp | 25 +- cpp/src/dual_simplex/right_looking_lu.cpp | 4 +- .../dual_simplex/simplex_solver_settings.hpp | 23 + cpp/src/dual_simplex/solution.hpp | 4 +- cpp/src/dual_simplex/solve.cpp | 73 +- cpp/src/dual_simplex/sparse_cholesky.cuh | 36 +- cpp/src/dual_simplex/sparse_matrix.cpp | 83 +- cpp/src/dual_simplex/sparse_matrix.hpp | 10 +- cpp/src/dual_simplex/sparse_vector.cpp | 50 +- cpp/src/dual_simplex/sparse_vector.hpp | 8 +- cpp/src/dual_simplex/types.hpp | 5 +- cpp/src/math_optimization/solver_settings.cu | 10 +- cpp/src/mip/diversity/diversity_manager.cu | 3 +- cpp/src/mip/diversity/lns/rins.cu | 8 +- cpp/src/mip/diversity/recombiners/sub_mip.cuh | 7 +- cpp/src/mip/presolve/third_party_presolve.cpp | 4 +- cpp/src/mip/solve.cu | 2 +- cpp/src/mip/solver.cu | 27 +- cpp/src/utilities/timer.hpp | 31 +- cpp/tests/mip/CMakeLists.txt | 5 +- cpp/tests/mip/cuts_test.cu | 168 + cpp/tests/mip/termination_test.cu | 4 +- .../linear_programming/test_python_API.py | 7 +- python/libcuopt/pyproject.toml | 1 - 47 files changed, 5178 insertions(+), 347 deletions(-) create mode 100644 cpp/src/dual_simplex/cuts.cpp create mode 100644 cpp/src/dual_simplex/cuts.hpp create mode 100644 cpp/tests/mip/cuts_test.cu diff --git a/cpp/include/cuopt/linear_programming/constants.h b/cpp/include/cuopt/linear_programming/constants.h index 4ee0224845..7caf7aeeb2 100644 --- a/cpp/include/cuopt/linear_programming/constants.h +++ b/cpp/include/cuopt/linear_programming/constants.h @@ -52,12 +52,20 @@ #define CUOPT_MIP_ABSOLUTE_TOLERANCE "mip_absolute_tolerance" #define CUOPT_MIP_RELATIVE_TOLERANCE "mip_relative_tolerance" #define CUOPT_MIP_INTEGRALITY_TOLERANCE "mip_integrality_tolerance" -#define CUOPT_MIP_BATCH_PDLP_STRONG_BRANCHING "mip_batch_pdlp_strong_branching" #define CUOPT_MIP_ABSOLUTE_GAP "mip_absolute_gap" #define CUOPT_MIP_RELATIVE_GAP "mip_relative_gap" #define CUOPT_MIP_HEURISTICS_ONLY "mip_heuristics_only" #define CUOPT_MIP_SCALING "mip_scaling" #define CUOPT_MIP_PRESOLVE "mip_presolve" +#define CUOPT_MIP_CUT_PASSES "mip_cut_passes" +#define CUOPT_MIP_MIXED_INTEGER_ROUNDING_CUTS "mip_mixed_integer_rounding_cuts" +#define CUOPT_MIP_MIXED_INTEGER_GOMORY_CUTS "mip_mixed_integer_gomory_cuts" +#define CUOPT_MIP_KNAPSACK_CUTS "mip_knapsack_cuts" +#define CUOPT_MIP_STRONG_CHVATAL_GOMORY_CUTS "mip_strong_chvatal_gomory_cuts" +#define CUOPT_MIP_REDUCED_COST_STRENGTHENING "mip_reduced_cost_strengthening" +#define CUOPT_MIP_CUT_CHANGE_THRESHOLD "mip_cut_change_threshold" +#define CUOPT_MIP_CUT_MIN_ORTHOGONALITY "mip_cut_min_orthogonality" +#define CUOPT_MIP_BATCH_PDLP_STRONG_BRANCHING "mip_batch_pdlp_strong_branching" #define CUOPT_SOLUTION_FILE "solution_file" #define CUOPT_NUM_CPU_THREADS "num_cpu_threads" #define CUOPT_NUM_GPUS "num_gpus" diff --git a/cpp/include/cuopt/linear_programming/mip/solver_settings.hpp b/cpp/include/cuopt/linear_programming/mip/solver_settings.hpp index 326d7f76ad..863e5d66d6 100644 --- a/cpp/include/cuopt/linear_programming/mip/solver_settings.hpp +++ b/cpp/include/cuopt/linear_programming/mip/solver_settings.hpp @@ -83,12 +83,22 @@ class mip_solver_settings_t { friend class problem_checking_t; tolerances_t tolerances; - f_t time_limit = std::numeric_limits::infinity(); - bool heuristics_only = false; - i_t num_cpu_threads = -1; // -1 means use default number of threads in branch and bound - i_t num_gpus = 1; + f_t time_limit = std::numeric_limits::infinity(); + i_t node_limit = std::numeric_limits::max(); + bool heuristics_only = false; + i_t num_cpu_threads = -1; // -1 means use default number of threads in branch and bound + i_t max_cut_passes = 10; // number of cut passes to make + i_t mir_cuts = -1; + i_t mixed_integer_gomory_cuts = -1; + i_t knapsack_cuts = -1; + i_t strong_chvatal_gomory_cuts = -1; + i_t reduced_cost_strengthening = -1; + f_t cut_change_threshold = 1e-3; + f_t cut_min_orthogonality = 0.5; i_t mip_batch_pdlp_strong_branching = 0; + i_t num_gpus = 1; bool log_to_console = true; + std::string log_file; std::string sol_file; std::string user_problem_file; diff --git a/cpp/src/dual_simplex/CMakeLists.txt b/cpp/src/dual_simplex/CMakeLists.txt index af1415fa9c..5405735043 100644 --- a/cpp/src/dual_simplex/CMakeLists.txt +++ b/cpp/src/dual_simplex/CMakeLists.txt @@ -10,6 +10,7 @@ set(DUAL_SIMPLEX_SRC_FILES ${CMAKE_CURRENT_SOURCE_DIR}/basis_updates.cpp ${CMAKE_CURRENT_SOURCE_DIR}/bound_flipping_ratio_test.cpp ${CMAKE_CURRENT_SOURCE_DIR}/branch_and_bound.cpp + ${CMAKE_CURRENT_SOURCE_DIR}/cuts.cpp ${CMAKE_CURRENT_SOURCE_DIR}/crossover.cpp ${CMAKE_CURRENT_SOURCE_DIR}/folding.cpp ${CMAKE_CURRENT_SOURCE_DIR}/initial_basis.cpp diff --git a/cpp/src/dual_simplex/barrier.cu b/cpp/src/dual_simplex/barrier.cu index a883232959..5eef97bb8a 100644 --- a/cpp/src/dual_simplex/barrier.cu +++ b/cpp/src/dual_simplex/barrier.cu @@ -681,7 +681,7 @@ class iteration_data_t { solve_status = chol->solve(U_col, M_col); if (solve_status != 0) { return solve_status; } if (settings_.concurrent_halt != nullptr && *settings_.concurrent_halt == 1) { - return -2; + return CONCURRENT_HALT_RETURN; } M.set_column(k, M_col); @@ -700,7 +700,7 @@ class iteration_data_t { AD_dense.transpose_multiply( 1.0, M.values.data() + k * M.m, 0.0, H.values.data() + k * H.m); if (settings_.concurrent_halt != nullptr && *settings_.concurrent_halt == 1) { - return -2; + return CONCURRENT_HALT_RETURN; } } @@ -1745,7 +1745,7 @@ int barrier_solver_t::initial_point(iteration_data_t& data) } else { status = data.chol->factorize(data.device_ADAT); } - if (status == -2) { return -2; } + if (status == CONCURRENT_HALT_RETURN) { return CONCURRENT_HALT_RETURN; } if (status != 0) { settings.log.printf("Initial factorization failed\n"); return -1; @@ -2309,7 +2309,7 @@ i_t barrier_solver_t::gpu_compute_search_direction(iteration_data_t::gpu_compute_search_direction(iteration_data_t& A, S_perm_inv); if (settings.concurrent_halt != nullptr && *settings.concurrent_halt == 1) { settings.log.printf("Concurrent halt\n"); - return -1; + return CONCURRENT_HALT_RETURN; } if (Srank != Sdim) { // Get the rank deficient columns @@ -582,7 +582,7 @@ i_t factorize_basis(const csc_matrix_t& A, } if (settings.concurrent_halt != nullptr && *settings.concurrent_halt == 1) { settings.log.printf("Concurrent halt\n"); - return -1; + return CONCURRENT_HALT_RETURN; } if (verbose) { printf("Right Lnz+Unz %d t %.3f\n", L.col_start[m] + U.col_start[m], toc(fact_start)); diff --git a/cpp/src/dual_simplex/basis_updates.cpp b/cpp/src/dual_simplex/basis_updates.cpp index 2c781a5156..1d94f41c7f 100644 --- a/cpp/src/dual_simplex/basis_updates.cpp +++ b/cpp/src/dual_simplex/basis_updates.cpp @@ -1108,6 +1108,217 @@ i_t basis_update_t::lower_triangular_multiply(const csc_matrix_t +i_t basis_update_mpf_t::append_cuts(const csr_matrix_t& cuts_basic) +{ + const i_t m = L0_.m; + + // Solve for U^T W^T = C_B^T + // We do this one row at a time of C_B + csc_matrix_t WT(m, cuts_basic.m, 0); + + i_t WT_nz = 0; + for (i_t k = 0; k < cuts_basic.m; k++) { + sparse_vector_t rhs(cuts_basic, k); + u_transpose_solve(rhs); + WT.col_start[k] = WT_nz; + for (i_t q = 0; q < rhs.i.size(); q++) { + WT.i.push_back(rhs.i[q]); + WT.x.push_back(rhs.x[q]); + WT_nz++; + } + } + WT.col_start[cuts_basic.m] = WT_nz; + +#ifdef CHECK_W + { + for (i_t k = 0; k < cuts_basic.m; k++) { + std::vector WT_col(m, 0.0); + WT.load_a_column(k, WT_col); + std::vector CBT_col(m, 0.0); + matrix_transpose_vector_multiply(U0_, 1.0, WT_col, 0.0, CBT_col); + sparse_vector_t CBT_col_sparse(cuts_basic, k); + std::vector CBT_col_dense(m); + CBT_col_sparse.to_dense(CBT_col_dense); + for (i_t h = 0; h < m; h++) { + if (std::abs(CBT_col_dense[h] - CBT_col[h]) > 1e-6) { + printf("W: col %d CBT_col_dense[%d] = %e CBT_col[%d] = %e\n", + k, + h, + CBT_col_dense[h], + h, + CBT_col[h]); + exit(1); + } + } + } + } +#endif + + csc_matrix_t V(cuts_basic.m, m, 0); + if (num_updates_ > 0) { + // W = V T_0 ... T_{num_updates_ - 1} + // or V = W T_{num_updates_ - 1}^{-1} ... T_0^{-1} + // or V^T = T_0^{-T} ... T_{num_updates_ - 1}^{-T} W^T + // We can compute V^T column by column so that we have + // V^T(:, h) = T_0^{-T} ... T_{num_updates_ - 1}^{-T} W^T(:, h) + // or + // V(h, :) = T_0^{-T} ... T_{num_updates_ - 1}^{-T} W^T(:, h) + // So we can form V row by row in CSR and then covert it to CSC + // for appending to L0 + + csr_matrix_t V_row(cuts_basic.m, m, 0); + i_t V_nz = 0; + const f_t zero_tol = 1e-13; + for (i_t h = 0; h < cuts_basic.m; h++) { + sparse_vector_t rhs(WT, h); + scatter_into_workspace(rhs); + i_t nz = rhs.i.size(); + for (i_t k = num_updates_ - 1; k >= 0; --k) { + // T_k^{-T} = ( I - v u^T/(1 + u^T v)) + // T_k^{-T} * b = b - v * (u^T * b) / (1 + u^T * v) = b - theta * v, theta = u^T b / mu + + const i_t u_col = 2 * k; + const i_t v_col = 2 * k + 1; + const f_t mu = mu_values_[k]; + + // dot = u^T * b + f_t dot = dot_product(u_col, xi_workspace_, x_workspace_); + const f_t theta = dot / mu; + if (std::abs(theta) > zero_tol) { + add_sparse_column(S_, v_col, -theta, xi_workspace_, nz, x_workspace_); + } + } + gather_into_sparse_vector(nz, rhs); + V_row.row_start[h] = V_nz; + for (i_t q = 0; q < rhs.i.size(); q++) { + V_row.j.push_back(rhs.i[q]); + V_row.x.push_back(rhs.x[q]); + V_nz++; + } + } + V_row.row_start[cuts_basic.m] = V_nz; + + V_row.to_compressed_col(V); + +#ifdef CHECK_V + csc_matrix_t CB_col(cuts_basic.m, m, 0); + cuts_basic.to_compressed_col(CB_col); + for (i_t k = 0; k < m; k++) { + std::vector U_col(m, 0.0); + U0_.load_a_column(k, U_col); + for (i_t h = num_updates_ - 1; h >= 0; --h) { + // T_h = ( I + u_h v_h^T) + // T_h * x = x + u_h * v_h^T * x = x + theta * u_h + const i_t u_col = 2 * h; + const i_t v_col = 2 * h + 1; + f_t theta = dot_product(v_col, U_col); + const i_t col_start = S_.col_start[u_col]; + const i_t col_end = S_.col_start[u_col + 1]; + for (i_t p = col_start; p < col_end; ++p) { + const i_t i = S_.i[p]; + U_col[i] += theta * S_.x[p]; + } + } + std::vector CB_column(cuts_basic.m, 0.0); + matrix_vector_multiply(V, 1.0, U_col, 0.0, CB_column); + std::vector CB_col_dense(cuts_basic.m); + CB_col.load_a_column(k, CB_col_dense); + for (i_t l = 0; l < cuts_basic.m; l++) { + if (std::abs(CB_col_dense[l] - CB_column[l]) > 1e-6) { + printf("V: col %d CB_col_dense[%d] = %e CB_column[%d] = %e\n", + k, + l, + CB_col_dense[l], + l, + CB_column[l]); + exit(1); + } + } + } +#endif + } else { + // W = V + WT.transpose(V); + } + + // Extend u_i, v_i for i = 0, ..., num_updates_ - 1 + S_.m += cuts_basic.m; + + // Adjust L and U + // L = [ L0 0 ] + // [ V I ] + + i_t V_nz = V.col_start[m]; + i_t L_nz = L0_.col_start[m]; + csc_matrix_t new_L(m + cuts_basic.m, m + cuts_basic.m, L_nz + V_nz + cuts_basic.m); + i_t predicted_nz = L_nz + V_nz + cuts_basic.m; + L_nz = 0; + for (i_t j = 0; j < m; ++j) { + new_L.col_start[j] = L_nz; + const i_t col_start = L0_.col_start[j]; + const i_t col_end = L0_.col_start[j + 1]; + for (i_t p = col_start; p < col_end; ++p) { + new_L.i[L_nz] = L0_.i[p]; + new_L.x[L_nz] = L0_.x[p]; + L_nz++; + } + const i_t V_col_start = V.col_start[j]; + const i_t V_col_end = V.col_start[j + 1]; + for (i_t p = V_col_start; p < V_col_end; ++p) { + new_L.i[L_nz] = V.i[p] + m; + new_L.x[L_nz] = V.x[p]; + L_nz++; + } + } + for (i_t j = m; j < m + cuts_basic.m; ++j) { + new_L.col_start[j] = L_nz; + new_L.i[L_nz] = j; + new_L.x[L_nz] = 1.0; + L_nz++; + } + new_L.col_start[m + cuts_basic.m] = L_nz; + assert(L_nz == predicted_nz); + + L0_ = new_L; + + // Adjust U + // U = [ U0 0 ] + // [ 0 I ] + + i_t U_nz = U0_.col_start[m]; + U0_.col_start.resize(m + cuts_basic.m + 1); + U0_.i.resize(U_nz + cuts_basic.m); + U0_.x.resize(U_nz + cuts_basic.m); + for (i_t k = m; k < m + cuts_basic.m; ++k) { + U0_.col_start[k] = U_nz; + U0_.i[U_nz] = k; + U0_.x[U_nz] = 1.0; + U_nz++; + } + U0_.col_start[m + cuts_basic.m] = U_nz; + U0_.n = m + cuts_basic.m; + U0_.m = m + cuts_basic.m; + + compute_transposes(); + + // Adjust row_permutation_ and inverse_row_permutation_ + row_permutation_.resize(m + cuts_basic.m); + inverse_row_permutation_.resize(m + cuts_basic.m); + for (i_t k = m; k < m + cuts_basic.m; ++k) { + row_permutation_[k] = k; + } + inverse_permutation(row_permutation_, inverse_row_permutation_); + + // Adjust workspace sizes + xi_workspace_.resize(2 * (m + cuts_basic.m), 0); + x_workspace_.resize(m + cuts_basic.m, 0.0); + + return 0; +} + template void basis_update_mpf_t::gather_into_sparse_vector(i_t nz, sparse_vector_t& out) const @@ -2057,16 +2268,18 @@ int basis_update_mpf_t::refactor_basis( if (L0_.m != A.m) { resize(A.m); } std::vector q; - if (factorize_basis(A, - settings, - basic_list, - L0_, - U0_, - row_permutation_, - inverse_row_permutation_, - q, - deficient, - slacks_needed) == -1) { + i_t status = factorize_basis(A, + settings, + basic_list, + L0_, + U0_, + row_permutation_, + inverse_row_permutation_, + q, + deficient, + slacks_needed); + if (status == CONCURRENT_HALT_RETURN) { return CONCURRENT_HALT_RETURN; } + if (status == -1) { settings.log.debug("Initial factorization failed\n"); basis_repair( A, settings, lower, upper, deficient, slacks_needed, basic_list, nonbasic_list, vstatus); @@ -2088,16 +2301,18 @@ int basis_update_mpf_t::refactor_basis( } #endif - if (factorize_basis(A, - settings, - basic_list, - L0_, - U0_, - row_permutation_, - inverse_row_permutation_, - q, - deficient, - slacks_needed) == -1) { + status = factorize_basis(A, + settings, + basic_list, + L0_, + U0_, + row_permutation_, + inverse_row_permutation_, + q, + deficient, + slacks_needed); + if (status == CONCURRENT_HALT_RETURN) { return CONCURRENT_HALT_RETURN; } + if (status == -1) { #ifdef CHECK_L_FACTOR if (L0_.check_matrix() == -1) { settings.log.printf("Bad L after basis repair\n"); } #endif diff --git a/cpp/src/dual_simplex/basis_updates.hpp b/cpp/src/dual_simplex/basis_updates.hpp index afd4f4c9ab..8eca3ba8a9 100644 --- a/cpp/src/dual_simplex/basis_updates.hpp +++ b/cpp/src/dual_simplex/basis_updates.hpp @@ -291,6 +291,8 @@ class basis_update_mpf_t { reset_stats(); } + i_t append_cuts(const csr_matrix_t& cuts_basic); + f_t estimate_solution_density(f_t rhs_nz, f_t sum, i_t& num_calls, bool& use_hypersparse) const { num_calls++; diff --git a/cpp/src/dual_simplex/bound_flipping_ratio_test.cpp b/cpp/src/dual_simplex/bound_flipping_ratio_test.cpp index 8e58c24f01..fac65b8140 100644 --- a/cpp/src/dual_simplex/bound_flipping_ratio_test.cpp +++ b/cpp/src/dual_simplex/bound_flipping_ratio_test.cpp @@ -1,6 +1,6 @@ /* clang-format off */ /* - * SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved. + * SPDX-FileCopyrightText: Copyright (c) 2025-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. * SPDX-License-Identifier: Apache-2.0 */ /* clang-format on */ @@ -84,10 +84,7 @@ i_t bound_flipping_ratio_test_t::single_pass(i_t start, step_length = min_val; nonbasic_entering = candidate; // this should be temporary, find root causes where the candidate is not filled - if (nonbasic_entering == -1) { - // -1,-2 and -3 are reserved for other things - return -4; - } + if (nonbasic_entering == -1) { return RATIO_TEST_NUMERICAL_ISSUES; } const i_t j = entering_index = nonbasic_list_[nonbasic_entering]; constexpr bool verbose = false; @@ -123,16 +120,16 @@ i_t bound_flipping_ratio_test_t::compute_step_length(f_t& step_length, if constexpr (verbose) { settings_.log.printf("Initial breakpoints %d\n", num_breakpoints); } if (num_breakpoints == 0) { nonbasic_entering = -1; - return -1; + return RATIO_TEST_NO_ENTERING_VARIABLE; } f_t slope = slope_; nonbasic_entering = -1; - i_t entering_index = -1; + i_t entering_index = RATIO_TEST_NO_ENTERING_VARIABLE; i_t k_idx = single_pass( 0, num_breakpoints, indicies, ratios, slope, step_length, nonbasic_entering, entering_index); - if (k_idx == -4) { return -4; } + if (k_idx == RATIO_TEST_NUMERICAL_ISSUES) { return RATIO_TEST_NUMERICAL_ISSUES; } bool continue_search = k_idx >= 0 && num_breakpoints > 1 && slope > 0.0; if (!continue_search) { if constexpr (0) { @@ -255,11 +252,11 @@ void bound_flipping_ratio_test_t::heap_passes(const std::vector& } if (toc(start_time_) > settings_.time_limit) { - entering_index = -2; + entering_index = RATIO_TEST_TIME_LIMIT; return; } if (settings_.concurrent_halt != nullptr && *settings_.concurrent_halt == 1) { - entering_index = -3; + entering_index = CONCURRENT_HALT_RETURN; return; } } diff --git a/cpp/src/dual_simplex/bound_flipping_ratio_test.hpp b/cpp/src/dual_simplex/bound_flipping_ratio_test.hpp index d3164c623d..51b00b1097 100644 --- a/cpp/src/dual_simplex/bound_flipping_ratio_test.hpp +++ b/cpp/src/dual_simplex/bound_flipping_ratio_test.hpp @@ -1,6 +1,6 @@ /* clang-format off */ /* - * SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved. + * SPDX-FileCopyrightText: Copyright (c) 2025-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. * SPDX-License-Identifier: Apache-2.0 */ /* clang-format on */ @@ -8,11 +8,17 @@ #include #include +#include #include namespace cuopt::linear_programming::dual_simplex { +#define RATIO_TEST_NO_ENTERING_VARIABLE -1 +#define RATIO_TEST_CONCURRENT_LIMIT CONCURRENT_HALT_RETURN // -2 +#define RATIO_TEST_TIME_LIMIT -3 +#define RATIO_TEST_NUMERICAL_ISSUES -4 + template class bound_flipping_ratio_test_t { public: diff --git a/cpp/src/dual_simplex/bounds_strengthening.cpp b/cpp/src/dual_simplex/bounds_strengthening.cpp index 4114e7e097..2b20940d29 100644 --- a/cpp/src/dual_simplex/bounds_strengthening.cpp +++ b/cpp/src/dual_simplex/bounds_strengthening.cpp @@ -59,8 +59,7 @@ bounds_strengthening_t::bounds_strengthening_t( const csr_matrix_t& Arow, const std::vector& row_sense, const std::vector& var_types) - : bounds_changed(problem.num_cols, false), - A(problem.A), + : A(problem.A), Arow(Arow), var_types(var_types), delta_min_activity(problem.num_rows), @@ -91,9 +90,10 @@ bounds_strengthening_t::bounds_strengthening_t( template bool bounds_strengthening_t::bounds_strengthening( + const simplex_solver_settings_t& settings, + const std::vector& bounds_changed, std::vector& lower_bounds, - std::vector& upper_bounds, - const simplex_solver_settings_t& settings) + std::vector& upper_bounds) { const i_t m = A.m; const i_t n = A.n; @@ -104,13 +104,13 @@ bool bounds_strengthening_t::bounds_strengthening( if (!bounds_changed.empty()) { std::fill(constraint_changed.begin(), constraint_changed.end(), false); - for (i_t i = 0; i < n; ++i) { - if (bounds_changed[i]) { - const i_t row_start = A.col_start[i]; - const i_t row_end = A.col_start[i + 1]; - for (i_t p = row_start; p < row_end; ++p) { - const i_t j = A.i[p]; - constraint_changed[j] = true; + for (i_t j = 0; j < n; ++j) { + if (bounds_changed[j]) { + const i_t col_start = A.col_start[j]; + const i_t col_end = A.col_start[j + 1]; + for (i_t p = col_start; p < col_end; ++p) { + const i_t i = A.i[p]; + constraint_changed[i] = true; } } } @@ -179,9 +179,9 @@ bool bounds_strengthening_t::bounds_strengthening( f_t new_lb = old_lb; f_t new_ub = old_ub; - const i_t row_start = A.col_start[k]; - const i_t row_end = A.col_start[k + 1]; - for (i_t p = row_start; p < row_end; ++p) { + const i_t col_start = A.col_start[k]; + const i_t col_end = A.col_start[k + 1]; + for (i_t p = col_start; p < col_end; ++p) { const i_t i = A.i[p]; if (!constraint_changed[i]) { continue; } @@ -210,13 +210,13 @@ bool bounds_strengthening_t::bounds_strengthening( new_lb = std::max(new_lb, lower_bounds[k]); new_ub = std::min(new_ub, upper_bounds[k]); - if (new_lb > new_ub + 1e-6) { + if (new_lb > new_ub + settings.primal_tol) { settings.log.debug( "Iter:: %d, Infeasible variable after update %d, %e > %e\n", iter, k, new_lb, new_ub); return false; } if (new_lb != old_lb || new_ub != old_ub) { - for (i_t p = row_start; p < row_end; ++p) { + for (i_t p = col_start; p < col_end; ++p) { const i_t i = A.i[p]; constraint_changed_next[i] = true; } @@ -225,8 +225,8 @@ bool bounds_strengthening_t::bounds_strengthening( lower[k] = std::min(new_lb, new_ub); upper[k] = std::max(new_lb, new_ub); - bool bounds_changed = lb_updated || ub_updated; - if (bounds_changed) { num_bounds_changed++; } + bool bounds_updated = lb_updated || ub_updated; + if (bounds_updated) { num_bounds_changed++; } } if (num_bounds_changed == 0) { break; } diff --git a/cpp/src/dual_simplex/bounds_strengthening.hpp b/cpp/src/dual_simplex/bounds_strengthening.hpp index e7e218b824..b811fb1c18 100644 --- a/cpp/src/dual_simplex/bounds_strengthening.hpp +++ b/cpp/src/dual_simplex/bounds_strengthening.hpp @@ -1,6 +1,6 @@ /* clang-format off */ /* - * SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved. + * SPDX-FileCopyrightText: Copyright (c) 2025-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. * SPDX-License-Identifier: Apache-2.0 */ /* clang-format on */ @@ -20,11 +20,12 @@ class bounds_strengthening_t { const std::vector& row_sense, const std::vector& var_types); - bool bounds_strengthening(std::vector& lower_bounds, - std::vector& upper_bounds, - const simplex_solver_settings_t& settings); - - std::vector bounds_changed; + // If bounds_changed is empty, all constraints are scanned for changes. + // Otherwise, bounds_changed must be a vector of length n, where n is the number of variables. + bool bounds_strengthening(const simplex_solver_settings_t& settings, + const std::vector& bounds_changed, + std::vector& lower_bounds, + std::vector& upper_bounds); private: const csc_matrix_t& A; diff --git a/cpp/src/dual_simplex/branch_and_bound.cpp b/cpp/src/dual_simplex/branch_and_bound.cpp index acdc9888a9..7f6e2c1921 100644 --- a/cpp/src/dual_simplex/branch_and_bound.cpp +++ b/cpp/src/dual_simplex/branch_and_bound.cpp @@ -7,8 +7,10 @@ #include +#include #include #include +#include #include #include #include @@ -122,9 +124,13 @@ bool check_guess(const lp_problem_t& original_lp, } template -void set_uninitialized_steepest_edge_norms(std::vector& edge_norms) +void set_uninitialized_steepest_edge_norms(const lp_problem_t& lp, + const std::vector& basic_list, + std::vector& edge_norms) { - for (i_t j = 0; j < edge_norms.size(); ++j) { + if (edge_norms.size() != lp.num_cols) { edge_norms.resize(lp.num_cols, -1.0); } + for (i_t k = 0; k < lp.num_rows; k++) { + const i_t j = basic_list[k]; if (edge_norms[j] <= 0.0) { edge_norms[j] = 1e-4; } } } @@ -225,7 +231,8 @@ inline char feasible_solution_symbol(bnb_worker_type_t type) template branch_and_bound_t::branch_and_bound_t( const user_problem_t& user_problem, - const simplex_solver_settings_t& solver_settings) + const simplex_solver_settings_t& solver_settings, + f_t start_time) : original_problem_(user_problem), settings_(solver_settings), original_lp_(user_problem.handle_ptr, 1, 1, 1), @@ -236,11 +243,36 @@ branch_and_bound_t::branch_and_bound_t( pc_(1), solver_status_(mip_status_t::UNSET) { - exploration_stats_.start_time = tic(); + exploration_stats_.start_time = start_time; +#ifdef PRINT_CONSTRAINT_MATRIX + settings_.log.printf("A"); + original_problem_.A.print_matrix(); +#endif + dualize_info_t dualize_info; convert_user_problem(original_problem_, settings_, original_lp_, new_slacks_, dualize_info); full_variable_types(original_problem_, original_lp_, var_types_); + // Check slack +#ifdef CHECK_SLACKS + assert(new_slacks_.size() == original_lp_.num_rows); + for (i_t slack : new_slacks_) { + const i_t col_start = original_lp_.A.col_start[slack]; + const i_t col_end = original_lp_.A.col_start[slack + 1]; + const i_t col_len = col_end - col_start; + if (col_len != 1) { + settings_.log.printf("Slack %d has %d nzs\n", slack, col_len); + assert(col_len == 1); + } + const i_t i = original_lp_.A.i[col_start]; + const f_t x = original_lp_.A.x[col_start]; + if (std::abs(x) != 1.0) { + settings_.log.printf("Slack %d row %d has non-unit coefficient %e\n", slack, i, x); + assert(std::abs(x) == 1.0); + } + } +#endif + upper_bound_ = inf; } @@ -267,7 +299,7 @@ void branch_and_bound_t::report_heuristic(f_t obj) std::string user_gap = user_mip_gap(user_obj, user_lower); settings_.log.printf( - "H %+13.6e %+10.6e %s %9.2f\n", + "H %+13.6e %+10.6e %s %9.2f\n", user_obj, user_lower, user_gap.c_str(), @@ -280,27 +312,90 @@ void branch_and_bound_t::report_heuristic(f_t obj) } template -void branch_and_bound_t::report(char symbol, f_t obj, f_t lower_bound, i_t node_depth) +void branch_and_bound_t::report( + char symbol, f_t obj, f_t lower_bound, i_t node_depth, i_t node_int_infeas) { update_user_bound(lower_bound); - i_t nodes_explored = exploration_stats_.nodes_explored; - i_t nodes_unexplored = exploration_stats_.nodes_unexplored; - f_t user_obj = compute_user_objective(original_lp_, obj); - f_t user_lower = compute_user_objective(original_lp_, lower_bound); - f_t iter_node = exploration_stats_.total_lp_iters / nodes_explored; - std::string user_gap = user_mip_gap(user_obj, user_lower); - settings_.log.printf("%c %10d %10lu %+13.6e %+10.6e %6d %7.1e %s %9.2f\n", + const i_t nodes_explored = exploration_stats_.nodes_explored; + const i_t nodes_unexplored = exploration_stats_.nodes_unexplored; + const f_t user_obj = compute_user_objective(original_lp_, obj); + const f_t user_lower = compute_user_objective(original_lp_, lower_bound); + const f_t iters = static_cast(exploration_stats_.total_lp_iters); + const f_t iter_node = nodes_explored > 0 ? iters / nodes_explored : iters; + const std::string user_gap = user_mip_gap(user_obj, user_lower); + settings_.log.printf("%c %10d %10lu %+13.6e %+10.6e %6d %6d %7.1e %s %9.2f\n", symbol, nodes_explored, nodes_unexplored, user_obj, user_lower, + node_int_infeas, node_depth, iter_node, user_gap.c_str(), toc(exploration_stats_.start_time)); } +template +i_t branch_and_bound_t::find_reduced_cost_fixings(f_t upper_bound, + std::vector& lower_bounds, + std::vector& upper_bounds) +{ + std::vector reduced_costs = root_relax_soln_.z; + lower_bounds = original_lp_.lower; + upper_bounds = original_lp_.upper; + std::vector bounds_changed(original_lp_.num_cols, false); + const f_t root_obj = compute_objective(original_lp_, root_relax_soln_.x); + const f_t threshold = 100.0 * settings_.integer_tol; + const f_t weaken = settings_.integer_tol; + const f_t fixed_tol = settings_.fixed_tol; + i_t num_improved = 0; + i_t num_fixed = 0; + i_t num_cols_to_check = reduced_costs.size(); // Reduced costs will be smaller than the original + // problem because we have added slacks for cuts + for (i_t j = 0; j < num_cols_to_check; j++) { + if (std::isfinite(reduced_costs[j]) && std::abs(reduced_costs[j]) > threshold) { + const f_t lower_j = original_lp_.lower[j]; + const f_t upper_j = original_lp_.upper[j]; + const f_t abs_gap = upper_bound - root_obj; + f_t reduced_cost_upper_bound = upper_j; + f_t reduced_cost_lower_bound = lower_j; + if (lower_j > -inf && reduced_costs[j] > 0) { + const f_t new_upper_bound = lower_j + abs_gap / reduced_costs[j]; + reduced_cost_upper_bound = var_types_[j] == variable_type_t::INTEGER + ? std::floor(new_upper_bound + weaken) + : new_upper_bound; + if (reduced_cost_upper_bound < upper_j && var_types_[j] == variable_type_t::INTEGER) { + num_improved++; + upper_bounds[j] = reduced_cost_upper_bound; + bounds_changed[j] = true; + } + } + if (upper_j < inf && reduced_costs[j] < 0) { + const f_t new_lower_bound = upper_j + abs_gap / reduced_costs[j]; + reduced_cost_lower_bound = var_types_[j] == variable_type_t::INTEGER + ? std::ceil(new_lower_bound - weaken) + : new_lower_bound; + if (reduced_cost_lower_bound > lower_j && var_types_[j] == variable_type_t::INTEGER) { + num_improved++; + lower_bounds[j] = reduced_cost_lower_bound; + bounds_changed[j] = true; + } + } + if (var_types_[j] == variable_type_t::INTEGER && + reduced_cost_upper_bound <= reduced_cost_lower_bound + fixed_tol) { + num_fixed++; + } + } + } + + if (num_fixed > 0 || num_improved > 0) { + settings_.log.printf( + "Reduced costs: Found %d improved bounds and %d fixed variables\n", num_improved, num_fixed); + } + return num_fixed; +} + template void branch_and_bound_t::update_user_bound(f_t lower_bound) { @@ -312,6 +407,7 @@ void branch_and_bound_t::update_user_bound(f_t lower_bound) template void branch_and_bound_t::set_new_solution(const std::vector& solution) { + mutex_original_lp_.lock(); if (solution.size() != original_problem_.num_cols) { settings_.log.printf( "Solution size mismatch %ld %d\n", solution.size(), original_problem_.num_cols); @@ -319,17 +415,29 @@ void branch_and_bound_t::set_new_solution(const std::vector& solu std::vector crushed_solution; crush_primal_solution( original_problem_, original_lp_, solution, new_slacks_, crushed_solution); - f_t obj = compute_objective(original_lp_, crushed_solution); + f_t obj = compute_objective(original_lp_, crushed_solution); + mutex_original_lp_.unlock(); bool is_feasible = false; bool attempt_repair = false; mutex_upper_.lock(); - if (obj < upper_bound_) { + f_t current_upper_bound = upper_bound_; + mutex_upper_.unlock(); + if (obj < current_upper_bound) { f_t primal_err; f_t bound_err; i_t num_fractional; + mutex_original_lp_.lock(); + if (crushed_solution.size() != original_lp_.num_cols) { + // original problem has been modified since the solution was crushed + // we need to re-crush the solution + crush_primal_solution( + original_problem_, original_lp_, solution, new_slacks_, crushed_solution); + } is_feasible = check_guess( original_lp_, settings_, var_types_, crushed_solution, primal_err, bound_err, num_fractional); - if (is_feasible) { + mutex_original_lp_.unlock(); + mutex_upper_.lock(); + if (is_feasible && obj < upper_bound_) { upper_bound_ = obj; incumbent_.set_incumbent_solution(obj, crushed_solution); } else { @@ -344,10 +452,11 @@ void branch_and_bound_t::set_new_solution(const std::vector& solu num_fractional); } } + mutex_upper_.unlock(); } else { settings_.log.debug("Solution objective not better than current upper_bound_. Not accepted.\n"); } - mutex_upper_.unlock(); + if (is_feasible) { report_heuristic(obj); } if (attempt_repair) { mutex_repair_.lock(); @@ -459,6 +568,35 @@ void branch_and_bound_t::repair_heuristic_solutions() } } +template +void branch_and_bound_t::set_solution_at_root(mip_solution_t& solution, + const cut_info_t& cut_info) +{ + mutex_upper_.lock(); + incumbent_.set_incumbent_solution(root_objective_, root_relax_soln_.x); + upper_bound_ = root_objective_; + mutex_upper_.unlock(); + + print_cut_info(settings_, cut_info); + + // We should be done here + uncrush_primal_solution(original_problem_, original_lp_, incumbent_.x, solution.x); + solution.objective = incumbent_.objective; + solution.lower_bound = root_objective_; + solution.nodes_explored = 0; + solution.simplex_iterations = root_relax_soln_.iterations; + settings_.log.printf("Optimal solution found at root node. Objective %.16e. Time %.2f.\n", + compute_user_objective(original_lp_, root_objective_), + toc(exploration_stats_.start_time)); + + if (settings_.solution_callback != nullptr) { + settings_.solution_callback(solution.x, solution.objective); + } + if (settings_.heuristic_preemption_callback != nullptr) { + settings_.heuristic_preemption_callback(); + } +} + template void branch_and_bound_t::set_final_solution(mip_solution_t& solution, f_t lower_bound) @@ -491,6 +629,9 @@ void branch_and_bound_t::set_final_solution(mip_solution_t& if (gap <= settings_.absolute_mip_gap_tol || gap_rel <= settings_.relative_mip_gap_tol) { solver_status_ = mip_status_t::OPTIMAL; +#ifdef CHECK_CUTS_AGAINST_SAVED_SOLUTION + if (settings_.sub_mip == 0) { write_solution_for_cut_verification(original_lp_, incumbent_.x); } +#endif if (gap > 0 && gap <= settings_.absolute_mip_gap_tol) { settings_.log.printf("Optimal solution found within absolute MIP gap tolerance (%.1e)\n", settings_.absolute_mip_gap_tol); @@ -542,7 +683,7 @@ void branch_and_bound_t::add_feasible_solution(f_t leaf_objective, if (leaf_objective < upper_bound_) { incumbent_.set_incumbent_solution(leaf_objective, leaf_solution); upper_bound_ = leaf_objective; - report(feasible_solution_symbol(thread_type), leaf_objective, get_lower_bound(), leaf_depth); + report(feasible_solution_symbol(thread_type), leaf_objective, get_lower_bound(), leaf_depth, 0); send_solution = true; } @@ -621,11 +762,40 @@ branch_variable_t branch_and_bound_t::variable_selection( } } +template +void branch_and_bound_t::initialize_diving_heuristics_settings( + std::vector& diving_strategies) +{ + diving_strategies.reserve(4); + + if (settings_.diving_settings.pseudocost_diving != 0) { + diving_strategies.push_back(bnb_worker_type_t::PSEUDOCOST_DIVING); + } + + if (settings_.diving_settings.line_search_diving != 0) { + diving_strategies.push_back(bnb_worker_type_t::LINE_SEARCH_DIVING); + } + + if (settings_.diving_settings.guided_diving != 0) { + diving_strategies.push_back(bnb_worker_type_t::GUIDED_DIVING); + } + + if (settings_.diving_settings.coefficient_diving != 0) { + diving_strategies.push_back(bnb_worker_type_t::COEFFICIENT_DIVING); + calculate_variable_locks(original_lp_, var_up_locks_, var_down_locks_); + } + + if (diving_strategies.empty()) { + settings_.log.printf("Warning: All diving heuristics are disabled!\n"); + } +} + template dual::status_t branch_and_bound_t::solve_node_lp( mip_node_t* node_ptr, lp_problem_t& leaf_problem, lp_solution_t& leaf_solution, + std::vector& leaf_edge_norms, basis_update_mpf_t& basis_factors, std::vector& basic_list, std::vector& nonbasic_list, @@ -637,6 +807,41 @@ dual::status_t branch_and_bound_t::solve_node_lp( bnb_stats_t& stats, logger_t& log) { +#ifdef DEBUG_BRANCHING + i_t num_integer_variables = 0; + for (i_t j = 0; j < original_lp_.num_cols; j++) { + if (var_types_[j] == variable_type_t::INTEGER) { num_integer_variables++; } + } + if (node_ptr->depth > num_integer_variables) { + std::vector branched_variables(original_lp_.num_cols, 0); + std::vector branched_lower(original_lp_.num_cols, std::numeric_limits::quiet_NaN()); + std::vector branched_upper(original_lp_.num_cols, std::numeric_limits::quiet_NaN()); + mip_node_t* parent = node_ptr->parent; + while (parent != nullptr) { + if (original_lp_.lower[parent->branch_var] != 0.0 || + original_lp_.upper[parent->branch_var] != 1.0) { + break; + } + if (branched_variables[parent->branch_var] == 1) { + printf( + "Variable %d already branched. Previous lower %e upper %e. Current lower %e upper %e.\n", + parent->branch_var, + branched_lower[parent->branch_var], + branched_upper[parent->branch_var], + parent->branch_var_lower, + parent->branch_var_upper); + } + branched_variables[parent->branch_var] = 1; + branched_lower[parent->branch_var] = parent->branch_var_lower; + branched_upper[parent->branch_var] = parent->branch_var_upper; + parent = parent->parent; + } + if (parent == nullptr) { + printf("Depth %d > num_integer_variables %d\n", node_ptr->depth, num_integer_variables); + } + } +#endif + std::vector& leaf_vstatus = node_ptr->vstatus; assert(leaf_vstatus.size() == leaf_problem.num_cols); @@ -678,29 +883,27 @@ dual::status_t branch_and_bound_t::solve_node_lp( #endif // Reset the bound_changed markers - std::fill(node_presolver.bounds_changed.begin(), node_presolver.bounds_changed.end(), false); + std::vector bounds_changed(original_lp_.num_cols, false); // Set the correct bounds for the leaf problem if (recompute_bounds_and_basis) { leaf_problem.lower = root_lower; leaf_problem.upper = root_upper; - node_ptr->get_variable_bounds( - leaf_problem.lower, leaf_problem.upper, node_presolver.bounds_changed); + node_ptr->get_variable_bounds(leaf_problem.lower, leaf_problem.upper, bounds_changed); } else { node_ptr->update_branched_variable_bounds( - leaf_problem.lower, leaf_problem.upper, node_presolver.bounds_changed); + leaf_problem.lower, leaf_problem.upper, bounds_changed); } - bool feasible = - node_presolver.bounds_strengthening(leaf_problem.lower, leaf_problem.upper, lp_settings); + bool feasible = node_presolver.bounds_strengthening( + lp_settings, bounds_changed, leaf_problem.lower, leaf_problem.upper); dual::status_t lp_status = dual::status_t::DUAL_UNBOUNDED; if (feasible) { - i_t node_iter = 0; - f_t lp_start_time = tic(); - std::vector leaf_edge_norms = edge_norms_; // = node.steepest_edge_norms; + i_t node_iter = 0; + f_t lp_start_time = tic(); lp_status = dual_phase2_with_advanced_basis(2, 0, @@ -717,7 +920,7 @@ dual::status_t branch_and_bound_t::solve_node_lp( leaf_edge_norms); if (lp_status == dual::status_t::NUMERICAL) { - log.printf("Numerical issue node %d. Resolving from scratch.\n", node_ptr->node_id); + log.debug("Numerical issue node %d. Resolving from scratch.\n", node_ptr->node_id); lp_status_t second_status = solve_linear_program_with_advanced_basis(leaf_problem, lp_start_time, lp_settings, @@ -748,6 +951,7 @@ std::pair branch_and_bound_t::upd search_tree_t& search_tree, lp_problem_t& leaf_problem, lp_solution_t& leaf_solution, + std::vector& leaf_edge_norms, bnb_worker_type_t thread_type, dual::status_t lp_status, logger_t& log) @@ -776,6 +980,23 @@ std::pair branch_and_bound_t::upd i_t leaf_num_fractional = fractional_variables(settings_, leaf_solution.x, var_types_, leaf_fractional); +#ifdef DEBUG_FRACTIONAL_FIXED + // Check if any of the fractional variables were fixed to their bounds + for (i_t j : leaf_fractional) { + if (leaf_problem.lower[j] == leaf_problem.upper[j]) { + printf( + "Node %d: Fixed variable %d has a fractional value %e. Lower %e upper %e. Variable " + "status %d\n", + node_ptr->node_id, + j, + leaf_solution.x[j], + leaf_problem.lower[j], + leaf_problem.upper[j], + leaf_vstatus[j]); + } + } +#endif + f_t leaf_objective = compute_objective(leaf_problem, leaf_solution.x); node_ptr->lower_bound = leaf_objective; search_tree.graphviz_node(log, node_ptr, "lower bound", leaf_objective); @@ -815,8 +1036,13 @@ std::pair branch_and_bound_t::upd pc_.obj_estimate(leaf_fractional, leaf_solution.x, node_ptr->lower_bound, pc_log); } - search_tree.branch( - node_ptr, branch_var, leaf_solution.x[branch_var], leaf_vstatus, leaf_problem, log); + search_tree.branch(node_ptr, + branch_var, + leaf_solution.x[branch_var], + leaf_num_fractional, + leaf_vstatus, + leaf_problem, + log); search_tree.update(node_ptr, node_status_t::HAS_CHILDREN); return {node_status_t::HAS_CHILDREN, round_dir}; @@ -877,7 +1103,7 @@ void branch_and_bound_t::exploration_ramp_up(mip_node_t* nod bool should_report = should_report_.exchange(false); if (should_report) { - report(' ', upper_bound, root_objective_, node->depth); + report(' ', upper_bound, root_objective_, node->depth, node->integer_infeasible); exploration_stats_.nodes_since_last_log = 0; exploration_stats_.last_log = tic(); should_report_ = true; @@ -900,9 +1126,11 @@ void branch_and_bound_t::exploration_ramp_up(mip_node_t* nod std::vector nonbasic_list; lp_solution_t leaf_solution(leaf_problem.num_rows, leaf_problem.num_cols); - dual::status_t lp_status = solve_node_lp(node, + std::vector leaf_edge_norms = edge_norms_; // = node.steepest_edge_norms; + dual::status_t lp_status = solve_node_lp(node, leaf_problem, leaf_solution, + leaf_edge_norms, basis_factors, basic_list, nonbasic_list, @@ -926,6 +1154,7 @@ void branch_and_bound_t::exploration_ramp_up(mip_node_t* nod search_tree_, leaf_problem, leaf_solution, + leaf_edge_norms, bnb_worker_type_t::BEST_FIRST, lp_status, settings_.log); @@ -1000,7 +1229,7 @@ void branch_and_bound_t::plunge_from(i_t task_id, abs_gap < 10 * settings_.absolute_mip_gap_tol) && time_since_last_log >= 1) || (time_since_last_log > 30) || now > settings_.time_limit) { - report(' ', upper_bound, get_lower_bound(), node_ptr->depth); + report(' ', upper_bound, get_lower_bound(), node_ptr->depth, node_ptr->integer_infeasible); exploration_stats_.last_log = tic(); exploration_stats_.nodes_since_last_log = 0; } @@ -1016,9 +1245,11 @@ void branch_and_bound_t::plunge_from(i_t task_id, } lp_solution_t leaf_solution(leaf_problem.num_rows, leaf_problem.num_cols); - dual::status_t lp_status = solve_node_lp(node_ptr, + std::vector leaf_edge_norms = edge_norms_; // = node.steepest_edge_norms; + dual::status_t lp_status = solve_node_lp(node_ptr, leaf_problem, leaf_solution, + leaf_edge_norms, basis_factors, basic_list, nonbasic_list, @@ -1045,6 +1276,7 @@ void branch_and_bound_t::plunge_from(i_t task_id, search_tree_, leaf_problem, leaf_solution, + leaf_edge_norms, bnb_worker_type_t::BEST_FIRST, lp_status, settings_.log); @@ -1187,9 +1419,11 @@ void branch_and_bound_t::dive_from(mip_node_t& start_node, if (dive_stats.nodes_explored > diving_node_limit) { break; } lp_solution_t leaf_solution(leaf_problem.num_rows, leaf_problem.num_cols); - dual::status_t lp_status = solve_node_lp(node_ptr, + std::vector leaf_edge_norms = edge_norms_; // = node.steepest_edge_norms; + dual::status_t lp_status = solve_node_lp(node_ptr, leaf_problem, leaf_solution, + leaf_edge_norms, basis_factors, basic_list, nonbasic_list, @@ -1210,9 +1444,15 @@ void branch_and_bound_t::dive_from(mip_node_t& start_node, ++dive_stats.nodes_explored; - auto [node_status, round_dir] = - update_tree(node_ptr, dive_tree, leaf_problem, leaf_solution, diving_type, lp_status, log); - recompute_bounds_and_basis = node_status != node_status_t::HAS_CHILDREN; + auto [node_status, round_dir] = update_tree(node_ptr, + dive_tree, + leaf_problem, + leaf_solution, + leaf_edge_norms, + diving_type, + lp_status, + log); + recompute_bounds_and_basis = node_status != node_status_t::HAS_CHILDREN; if (node_status == node_status_t::HAS_CHILDREN) { if (round_dir == rounding_direction_t::UP) { @@ -1239,6 +1479,7 @@ void branch_and_bound_t::diving_thread(bnb_worker_type_t diving_type) lp_problem_t leaf_problem = original_lp_; std::vector row_sense; bounds_strengthening_t node_presolver(leaf_problem, Arow_, row_sense, var_types_); + std::vector bounds_changed(original_lp_.num_cols, false); const i_t m = leaf_problem.num_rows; basis_update_mpf_t basis_factors(m, settings_.refactor_frequency); @@ -1254,7 +1495,7 @@ void branch_and_bound_t::diving_thread(bnb_worker_type_t diving_type) if (reset_starting_bounds) { start_lower = original_lp_.lower; start_upper = original_lp_.upper; - std::fill(node_presolver.bounds_changed.begin(), node_presolver.bounds_changed.end(), false); + std::fill(bounds_changed.begin(), bounds_changed.end(), false); reset_starting_bounds = false; } @@ -1268,8 +1509,7 @@ void branch_and_bound_t::diving_thread(bnb_worker_type_t diving_type) std::optional> start_node = std::nullopt; if (node_ptr.has_value()) { - node_ptr.value()->get_variable_bounds( - start_lower, start_upper, node_presolver.bounds_changed); + node_ptr.value()->get_variable_bounds(start_lower, start_upper, bounds_changed); start_node = node_ptr.value()->detach_copy(); } node_queue_.unlock(); @@ -1278,7 +1518,8 @@ void branch_and_bound_t::diving_thread(bnb_worker_type_t diving_type) reset_starting_bounds = true; if (upper_bound_ < start_node->lower_bound) { continue; } - bool is_feasible = node_presolver.bounds_strengthening(start_lower, start_upper, settings_); + bool is_feasible = + node_presolver.bounds_strengthening(settings_, bounds_changed, start_lower, start_upper); if (!is_feasible) { continue; } dive_from(start_node.value(), @@ -1296,7 +1537,13 @@ void branch_and_bound_t::diving_thread(bnb_worker_type_t diving_type) template lp_status_t branch_and_bound_t::solve_root_relaxation( - simplex_solver_settings_t const& lp_settings) + simplex_solver_settings_t const& lp_settings, + lp_solution_t& root_relax_soln, + std::vector& root_vstatus, + basis_update_mpf_t& basis_update, + std::vector& basic_list, + std::vector& nonbasic_list, + std::vector& edge_norms) { f_t start_time = tic(); f_t user_objective = 0; @@ -1307,13 +1554,16 @@ lp_status_t branch_and_bound_t::solve_root_relaxation( lp_status_t root_status; std::future root_status_future; root_status_future = std::async(std::launch::async, - &solve_linear_program_advanced, + &solve_linear_program_with_advanced_basis, std::ref(original_lp_), exploration_stats_.start_time, std::ref(lp_settings), - std::ref(root_relax_soln_), - std::ref(root_vstatus_), - std::ref(edge_norms_)); + std::ref(root_relax_soln), + std::ref(basis_update), + std::ref(basic_list), + std::ref(nonbasic_list), + std::ref(root_vstatus), + std::ref(edge_norms)); // Wait for the root relaxation solution to be sent by the diversity manager or dual simplex // to finish while (!root_crossover_solution_set_.load(std::memory_order_acquire) && @@ -1355,17 +1605,55 @@ lp_status_t branch_and_bound_t::solve_root_relaxation( // Check if crossover was stopped by dual simplex if (crossover_status == crossover_status_t::OPTIMAL) { - set_root_concurrent_halt(1); // Stop dual simplex - root_status = root_status_future.get(); - + set_root_concurrent_halt(1); // Stop dual simplex + root_status = root_status_future.get(); // Wait for dual simplex to finish + set_root_concurrent_halt(0); // Clear the concurrent halt flag // Override the root relaxation solution with the crossover solution - root_relax_soln_ = root_crossover_soln_; - root_vstatus_ = crossover_vstatus_; - root_status = lp_status_t::OPTIMAL; - user_objective = root_crossover_soln_.user_objective; - iter = root_crossover_soln_.iterations; - solver_name = "Barrier/PDLP and Crossover"; + root_relax_soln = root_crossover_soln_; + root_vstatus = crossover_vstatus_; + root_status = lp_status_t::OPTIMAL; + basic_list.clear(); + nonbasic_list.reserve(original_lp_.num_cols - original_lp_.num_rows); + nonbasic_list.clear(); + // Get the basic list and nonbasic list from the vstatus + for (i_t j = 0; j < original_lp_.num_cols; j++) { + if (crossover_vstatus_[j] == variable_status_t::BASIC) { + basic_list.push_back(j); + } else { + nonbasic_list.push_back(j); + } + } + if (basic_list.size() != original_lp_.num_rows) { + settings_.log.printf( + "basic_list size %d != m %d\n", basic_list.size(), original_lp_.num_rows); + assert(basic_list.size() == original_lp_.num_rows); + } + if (nonbasic_list.size() != original_lp_.num_cols - original_lp_.num_rows) { + settings_.log.printf("nonbasic_list size %d != n - m %d\n", + nonbasic_list.size(), + original_lp_.num_cols - original_lp_.num_rows); + assert(nonbasic_list.size() == original_lp_.num_cols - original_lp_.num_rows); + } + // Populate the basis_update from the crossover vstatus + i_t refactor_status = basis_update.refactor_basis(original_lp_.A, + root_crossover_settings, + original_lp_.lower, + original_lp_.upper, + basic_list, + nonbasic_list, + crossover_vstatus_); + if (refactor_status != 0) { + settings_.log.printf("Failed to refactor basis. %d deficient columns.\n", refactor_status); + assert(refactor_status == 0); + root_status = lp_status_t::NUMERICAL_ISSUES; + } + // Set the edge norms to a default value + edge_norms.resize(original_lp_.num_cols, -1.0); + set_uninitialized_steepest_edge_norms(original_lp_, basic_list, edge_norms); + user_objective = root_crossover_soln_.user_objective; + iter = root_crossover_soln_.iterations; + solver_name = "Barrier/PDLP and Crossover"; } else { root_status = root_status_future.get(); user_objective = root_relax_soln_.user_objective; @@ -1410,28 +1698,7 @@ mip_status_t branch_and_bound_t::solve(mip_solution_t& solut original_lp_.A.to_compressed_row(Arow_); std::vector diving_strategies; - diving_strategies.reserve(4); - - if (settings_.diving_settings.pseudocost_diving != 0) { - diving_strategies.push_back(bnb_worker_type_t::PSEUDOCOST_DIVING); - } - - if (settings_.diving_settings.line_search_diving != 0) { - diving_strategies.push_back(bnb_worker_type_t::LINE_SEARCH_DIVING); - } - - if (settings_.diving_settings.guided_diving != 0) { - diving_strategies.push_back(bnb_worker_type_t::GUIDED_DIVING); - } - - if (settings_.diving_settings.coefficient_diving != 0) { - diving_strategies.push_back(bnb_worker_type_t::COEFFICIENT_DIVING); - calculate_variable_locks(original_lp_, var_up_locks_, var_down_locks_); - } - - if (diving_strategies.empty()) { - settings_.log.printf("Warning: All diving heuristics are disabled!\n"); - } + initialize_diving_heuristics_settings(diving_strategies); if (guess_.size() != 0) { std::vector crushed_guess; @@ -1452,25 +1719,37 @@ mip_status_t branch_and_bound_t::solve(mip_solution_t& solut root_relax_soln_.resize(original_lp_.num_rows, original_lp_.num_cols); - lp_status_t root_status; + i_t original_rows = original_lp_.num_rows; simplex_solver_settings_t lp_settings = settings_; lp_settings.inside_mip = 1; + lp_settings.scale_columns = false; lp_settings.concurrent_halt = get_root_concurrent_halt(); - // RINS/SUBMIP path + std::vector basic_list(original_lp_.num_rows); + std::vector nonbasic_list; + basis_update_mpf_t basis_update(original_lp_.num_rows, settings_.refactor_frequency); + lp_status_t root_status; if (!enable_concurrent_lp_root_solve()) { + // RINS/SUBMIP path settings_.log.printf("\nSolving LP root relaxation with dual simplex\n"); - root_status = solve_linear_program_advanced(original_lp_, - exploration_stats_.start_time, - lp_settings, - root_relax_soln_, - root_vstatus_, - edge_norms_); - + root_status = solve_linear_program_with_advanced_basis(original_lp_, + exploration_stats_.start_time, + lp_settings, + root_relax_soln_, + basis_update, + basic_list, + nonbasic_list, + root_vstatus_, + edge_norms_); } else { settings_.log.printf("\nSolving LP root relaxation in concurrent mode\n"); - root_status = solve_root_relaxation(lp_settings); + root_status = solve_root_relaxation(lp_settings, + root_relax_soln_, + root_vstatus_, + basis_update, + basic_list, + nonbasic_list, + edge_norms_); } - exploration_stats_.total_lp_iters = root_relax_soln_.iterations; exploration_stats_.total_lp_solve_time = toc(exploration_stats_.start_time); @@ -1491,15 +1770,19 @@ mip_status_t branch_and_bound_t::solve(mip_solution_t& solut } return mip_status_t::UNBOUNDED; } - if (root_status == lp_status_t::TIME_LIMIT) { solver_status_ = mip_status_t::TIME_LIMIT; set_final_solution(solution, -inf); return solver_status_; } + if (root_status == lp_status_t::NUMERICAL_ISSUES) { + solver_status_ = mip_status_t::NUMERICAL; + set_final_solution(solution, -inf); + return solver_status_; + } assert(root_vstatus_.size() == original_lp_.num_cols); - set_uninitialized_steepest_edge_norms(edge_norms_); + set_uninitialized_steepest_edge_norms(original_lp_, basic_list, edge_norms_); root_objective_ = compute_objective(original_lp_, root_relax_soln_.x); local_lower_bounds_.assign(settings_.num_bfs_workers, root_objective_); @@ -1520,33 +1803,280 @@ mip_status_t branch_and_bound_t::solve(mip_solution_t& solut } std::vector fractional; - const i_t num_fractional = - fractional_variables(settings_, root_relax_soln_.x, var_types_, fractional); + i_t num_fractional = fractional_variables(settings_, root_relax_soln_.x, var_types_, fractional); + cut_info_t cut_info; if (num_fractional == 0) { - mutex_upper_.lock(); - incumbent_.set_incumbent_solution(root_objective_, root_relax_soln_.x); - upper_bound_ = root_objective_; - mutex_upper_.unlock(); - // We should be done here - uncrush_primal_solution(original_problem_, original_lp_, incumbent_.x, solution.x); - solution.objective = incumbent_.objective; - solution.lower_bound = root_objective_; - solution.nodes_explored = 0; - solution.simplex_iterations = root_relax_soln_.iterations; - settings_.log.printf("Optimal solution found at root node. Objective %.16e. Time %.2f.\n", - compute_user_objective(original_lp_, root_objective_), - toc(exploration_stats_.start_time)); + set_solution_at_root(solution, cut_info); + return mip_status_t::OPTIMAL; + } - if (settings_.solution_callback != nullptr) { - settings_.solution_callback(solution.x, solution.objective); - } - if (settings_.heuristic_preemption_callback != nullptr) { - settings_.heuristic_preemption_callback(); + is_running = true; + lower_bound_ceiling_ = inf; + + if (num_fractional != 0 && settings_.max_cut_passes > 0) { + settings_.log.printf( + " | Explored | Unexplored | Objective | Bound | IntInf | Depth | Iter/Node | " + "Gap " + "| Time |\n"); + } + + cut_pool_t cut_pool(original_lp_.num_cols, settings_); + cut_generation_t cut_generation( + cut_pool, original_lp_, settings_, Arow_, new_slacks_, var_types_); + + std::vector saved_solution; +#ifdef CHECK_CUTS_AGAINST_SAVED_SOLUTION + read_saved_solution_for_cut_verification(original_lp_, settings_, saved_solution); +#endif + + f_t last_upper_bound = std::numeric_limits::infinity(); + f_t last_objective = root_objective_; + f_t root_relax_objective = root_objective_; + + i_t cut_pool_size = 0; + for (i_t cut_pass = 0; cut_pass < settings_.max_cut_passes; cut_pass++) { + if (num_fractional == 0) { + set_solution_at_root(solution, cut_info); + return mip_status_t::OPTIMAL; + } else { +#ifdef PRINT_FRACTIONAL_INFO + settings_.log.printf( + "Found %d fractional variables on cut pass %d\n", num_fractional, cut_pass); + for (i_t j : fractional) { + settings_.log.printf("Fractional variable %d lower %e value %e upper %e\n", + j, + original_lp_.lower[j], + root_relax_soln_.x[j], + original_lp_.upper[j]); + } +#endif + + // Generate cuts and add them to the cut pool + f_t cut_start_time = tic(); + cut_generation.generate_cuts(original_lp_, + settings_, + Arow_, + new_slacks_, + var_types_, + basis_update, + root_relax_soln_.x, + basic_list, + nonbasic_list); + f_t cut_generation_time = toc(cut_start_time); + if (cut_generation_time > 1.0) { + settings_.log.debug("Cut generation time %.2f seconds\n", cut_generation_time); + } + // Score the cuts + f_t score_start_time = tic(); + cut_pool.score_cuts(root_relax_soln_.x); + f_t score_time = toc(score_start_time); + if (score_time > 1.0) { settings_.log.debug("Cut scoring time %.2f seconds\n", score_time); } + // Get the best cuts from the cut pool + csr_matrix_t cuts_to_add(0, original_lp_.num_cols, 0); + std::vector cut_rhs; + std::vector cut_types; + i_t num_cuts = cut_pool.get_best_cuts(cuts_to_add, cut_rhs, cut_types); + if (num_cuts == 0) { break; } + cut_info.record_cut_types(cut_types); +#ifdef PRINT_CUT_POOL_TYPES + cut_pool.print_cutpool_types(); + print_cut_types("In LP ", cut_types, settings_); + printf("Cut pool size: %d\n", cut_pool.pool_size()); +#endif + +#ifdef CHECK_CUT_MATRIX + if (cuts_to_add.check_matrix() != 0) { + settings_.log.printf("Bad cuts matrix\n"); + for (i_t i = 0; i < static_cast(cut_types.size()); ++i) { + settings_.log.printf("row %d cut type %d\n", i, cut_types[i]); + } + return mip_status_t::NUMERICAL; + } +#endif + // Check against saved solution +#ifdef CHECK_CUTS_AGAINST_SAVED_SOLUTION + verify_cuts_against_saved_solution(cuts_to_add, cut_rhs, saved_solution); +#endif + cut_pool_size = cut_pool.pool_size(); + + // Resolve the LP with the new cuts + settings_.log.debug( + "Solving LP with %d cuts (%d cut nonzeros). Cuts in pool %d. Total constraints %d\n", + num_cuts, + cuts_to_add.row_start[cuts_to_add.m], + cut_pool.pool_size(), + cuts_to_add.m + original_lp_.num_rows); + lp_settings.log.log = false; + + f_t add_cuts_start_time = tic(); + mutex_original_lp_.lock(); + i_t add_cuts_status = add_cuts(settings_, + cuts_to_add, + cut_rhs, + original_lp_, + new_slacks_, + root_relax_soln_, + basis_update, + basic_list, + nonbasic_list, + root_vstatus_, + edge_norms_); + var_types_.resize(original_lp_.num_cols, variable_type_t::CONTINUOUS); + mutex_original_lp_.unlock(); + f_t add_cuts_time = toc(add_cuts_start_time); + if (add_cuts_time > 1.0) { + settings_.log.debug("Add cuts time %.2f seconds\n", add_cuts_time); + } + if (add_cuts_status != 0) { + settings_.log.printf("Failed to add cuts\n"); + return mip_status_t::NUMERICAL; + } + + if (settings_.reduced_cost_strengthening >= 1 && upper_bound_.load() < last_upper_bound) { + mutex_upper_.lock(); + last_upper_bound = upper_bound_.load(); + std::vector lower_bounds; + std::vector upper_bounds; + find_reduced_cost_fixings(upper_bound_.load(), lower_bounds, upper_bounds); + mutex_upper_.unlock(); + mutex_original_lp_.lock(); + original_lp_.lower = lower_bounds; + original_lp_.upper = upper_bounds; + mutex_original_lp_.unlock(); + } + + // Try to do bound strengthening + std::vector bounds_changed(original_lp_.num_cols, true); + std::vector row_sense; +#ifdef CHECK_MATRICES + settings_.log.printf("Before A check\n"); + original_lp_.A.check_matrix(); +#endif + original_lp_.A.to_compressed_row(Arow_); + + f_t node_presolve_start_time = tic(); + bounds_strengthening_t node_presolve(original_lp_, Arow_, row_sense, var_types_); + std::vector new_lower = original_lp_.lower; + std::vector new_upper = original_lp_.upper; + bool feasible = + node_presolve.bounds_strengthening(settings_, bounds_changed, new_lower, new_upper); + mutex_original_lp_.lock(); + original_lp_.lower = new_lower; + original_lp_.upper = new_upper; + mutex_original_lp_.unlock(); + f_t node_presolve_time = toc(node_presolve_start_time); + if (node_presolve_time > 1.0) { + settings_.log.debug("Node presolve time %.2f seconds\n", node_presolve_time); + } + if (!feasible) { + settings_.log.printf("Bound strengthening detected infeasibility\n"); + return mip_status_t::INFEASIBLE; + } + + i_t iter = 0; + bool initialize_basis = false; + lp_settings.concurrent_halt = NULL; + f_t dual_phase2_start_time = tic(); + dual::status_t cut_status = dual_phase2_with_advanced_basis(2, + 0, + initialize_basis, + exploration_stats_.start_time, + original_lp_, + lp_settings, + root_vstatus_, + basis_update, + basic_list, + nonbasic_list, + root_relax_soln_, + iter, + edge_norms_); + f_t dual_phase2_time = toc(dual_phase2_start_time); + if (dual_phase2_time > 1.0) { + settings_.log.debug("Dual phase2 time %.2f seconds\n", dual_phase2_time); + } + if (cut_status == dual::status_t::TIME_LIMIT) { + solver_status_ = mip_status_t::TIME_LIMIT; + set_final_solution(solution, root_objective_); + return solver_status_; + } + + if (cut_status != dual::status_t::OPTIMAL) { + settings_.log.printf("Cut status %s\n", dual::status_to_string(cut_status).c_str()); + return mip_status_t::NUMERICAL; + } + exploration_stats_.total_lp_iters += root_relax_soln_.iterations; + root_objective_ = compute_objective(original_lp_, root_relax_soln_.x); + + local_lower_bounds_.assign(settings_.num_bfs_workers, root_objective_); + + f_t remove_cuts_start_time = tic(); + mutex_original_lp_.lock(); + remove_cuts(original_lp_, + settings_, + Arow_, + new_slacks_, + original_rows, + var_types_, + root_vstatus_, + edge_norms_, + root_relax_soln_.x, + root_relax_soln_.y, + root_relax_soln_.z, + basic_list, + nonbasic_list, + basis_update); + mutex_original_lp_.unlock(); + f_t remove_cuts_time = toc(remove_cuts_start_time); + if (remove_cuts_time > 1.0) { + settings_.log.debug("Remove cuts time %.2f seconds\n", remove_cuts_time); + } + fractional.clear(); + num_fractional = fractional_variables(settings_, root_relax_soln_.x, var_types_, fractional); + + if (num_fractional == 0) { + upper_bound_ = root_objective_; + mutex_upper_.lock(); + incumbent_.set_incumbent_solution(root_objective_, root_relax_soln_.x); + mutex_upper_.unlock(); + } + f_t obj = upper_bound_.load(); + report(' ', obj, root_objective_, 0, num_fractional); + + f_t rel_gap = user_relative_gap(original_lp_, upper_bound_.load(), root_objective_); + f_t abs_gap = upper_bound_.load() - root_objective_; + if (rel_gap < settings_.relative_mip_gap_tol || abs_gap < settings_.absolute_mip_gap_tol) { + set_solution_at_root(solution, cut_info); + set_final_solution(solution, root_objective_); + return mip_status_t::OPTIMAL; + } + + f_t change_in_objective = root_objective_ - last_objective; + const f_t factor = settings_.cut_change_threshold; + const f_t min_objective = 1e-3; + if (change_in_objective <= factor * std::max(min_objective, std::abs(root_relax_objective))) { + settings_.log.debug( + "Change in objective %.16e is less than 1e-3 of root relax objective %.16e\n", + change_in_objective, + root_relax_objective); + break; + } + last_objective = root_objective_; } - return mip_status_t::OPTIMAL; } + print_cut_info(settings_, cut_info); + + if (cut_info.has_cuts()) { + settings_.log.printf("Cut pool size : %d\n", cut_pool_size); + settings_.log.printf("Size with cuts : %d constraints, %d variables, %d nonzeros\n", + original_lp_.num_rows, + original_lp_.num_cols, + original_lp_.A.col_start[original_lp_.A.n]); + } + + set_uninitialized_steepest_edge_norms(original_lp_, basic_list, edge_norms_); + pc_.resize(original_lp_.num_cols); strong_branching(original_problem_, original_lp_, @@ -1566,6 +2096,50 @@ mip_status_t branch_and_bound_t::solve(mip_solution_t& solut return solver_status_; } + if (settings_.reduced_cost_strengthening >= 2 && upper_bound_.load() < last_upper_bound) { + std::vector lower_bounds; + std::vector upper_bounds; + i_t num_fixed = find_reduced_cost_fixings(upper_bound_.load(), lower_bounds, upper_bounds); + if (num_fixed > 0) { + std::vector bounds_changed(original_lp_.num_cols, true); + std::vector row_sense; + + bounds_strengthening_t node_presolve(original_lp_, Arow_, row_sense, var_types_); + + mutex_original_lp_.lock(); + original_lp_.lower = lower_bounds; + original_lp_.upper = upper_bounds; + bool feasible = node_presolve.bounds_strengthening( + settings_, bounds_changed, original_lp_.lower, original_lp_.upper); + mutex_original_lp_.unlock(); + if (!feasible) { + settings_.log.printf("Bound strengthening failed\n"); + return mip_status_t::NUMERICAL; // We had a feasible integer solution, but bound + // strengthening thinks we are infeasible. + } + // Go through and check the fractional variables and remove any that are now fixed to their + // bounds + std::vector to_remove(fractional.size(), 0); + i_t num_to_remove = 0; + for (i_t k = 0; k < fractional.size(); k++) { + const i_t j = fractional[k]; + if (std::abs(original_lp_.upper[j] - original_lp_.lower[j]) < settings_.fixed_tol) { + to_remove[k] = 1; + num_to_remove++; + } + } + if (num_to_remove > 0) { + std::vector new_fractional; + new_fractional.reserve(fractional.size() - num_to_remove); + for (i_t k = 0; k < fractional.size(); k++) { + if (!to_remove[k]) { new_fractional.push_back(fractional[k]); } + } + fractional = new_fractional; + num_fractional = fractional.size(); + } + } + } + // Choose variable to branch on i_t branch_var = pc_.variable_selection(fractional, root_relax_soln_.x, log); @@ -1575,6 +2149,7 @@ mip_status_t branch_and_bound_t::solve(mip_solution_t& solut search_tree_.branch(&search_tree_.root, branch_var, root_relax_soln_.x[branch_var], + num_fractional, root_vstatus_, original_lp_, log); @@ -1589,14 +2164,13 @@ mip_status_t branch_and_bound_t::solve(mip_solution_t& solut exploration_stats_.nodes_since_last_log = 0; exploration_stats_.last_log = tic(); active_subtrees_ = 0; - is_running = true; lower_bound_ceiling_ = inf; should_report_ = true; settings_.log.printf( - " | Explored | Unexplored | Objective | Bound | Depth | Iter/Node | Gap " + " | Explored | Unexplored | Objective | Bound | IntInf | Depth | Iter/Node | " + "Gap " "| Time |\n"); - #pragma omp parallel num_threads(settings_.num_threads) { #pragma omp master diff --git a/cpp/src/dual_simplex/branch_and_bound.hpp b/cpp/src/dual_simplex/branch_and_bound.hpp index 19621b8896..6db45e1531 100644 --- a/cpp/src/dual_simplex/branch_and_bound.hpp +++ b/cpp/src/dual_simplex/branch_and_bound.hpp @@ -7,6 +7,7 @@ #pragma once +#include #include #include #include @@ -72,7 +73,8 @@ template class branch_and_bound_t { public: branch_and_bound_t(const user_problem_t& user_problem, - const simplex_solver_settings_t& solver_settings); + const simplex_solver_settings_t& solver_settings, + f_t start_time); // Set an initial guess based on the user_problem. This should be called before solve. void set_initial_guess(const std::vector& user_guess) { guess_ = user_guess; } @@ -117,7 +119,17 @@ class branch_and_bound_t { bool enable_concurrent_lp_root_solve() const { return enable_concurrent_lp_root_solve_; } std::atomic* get_root_concurrent_halt() { return &root_concurrent_halt_; } void set_root_concurrent_halt(int value) { root_concurrent_halt_ = value; } - lp_status_t solve_root_relaxation(simplex_solver_settings_t const& lp_settings); + lp_status_t solve_root_relaxation(simplex_solver_settings_t const& lp_settings, + lp_solution_t& root_relax_soln, + std::vector& root_vstatus, + basis_update_mpf_t& basis_update, + std::vector& basic_list, + std::vector& nonbasic_list, + std::vector& edge_norms); + + i_t find_reduced_cost_fixings(f_t upper_bound, + std::vector& lower_bounds, + std::vector& upper_bounds); // The main entry routine. Returns the solver status and populates solution with the incumbent. mip_status_t solve(mip_solution_t& solution); @@ -144,6 +156,13 @@ class branch_and_bound_t { // Local lower bounds for each thread std::vector> local_lower_bounds_; + // Mutex for the original LP + // The heuristics threads look at the original LP. But the main thread modifies the + // size of the original LP by adding slacks for cuts. Heuristic threads should lock + // this mutex when accessing the original LP. The main thread should lock this mutex + // when modifying the original LP. + omp_mutex_t mutex_original_lp_; + // Mutex for upper bound omp_mutex_t mutex_upper_; @@ -196,7 +215,11 @@ class branch_and_bound_t { std::function user_bound_callback_; void report_heuristic(f_t obj); - void report(char symbol, f_t obj, f_t lower_bound, i_t node_depth); + void report(char symbol, f_t obj, f_t lower_bound, i_t node_depth, i_t node_int_infeas); + + // Set the solution when found at the root node + void set_solution_at_root(mip_solution_t& solution, + const cut_info_t& cut_info); void update_user_bound(f_t lower_bound); // Set the final solution. @@ -212,6 +235,9 @@ class branch_and_bound_t { // Repairs low-quality solutions from the heuristics, if it is applicable. void repair_heuristic_solutions(); + // Initialize diving heuristics settings + void initialize_diving_heuristics_settings(std::vector& diving_strategies); + // Ramp-up phase of the solver, where we greedily expand the tree until // there is enough unexplored nodes. This is done recursively using OpenMP tasks. void exploration_ramp_up(mip_node_t* node, i_t initial_heap_size); @@ -251,6 +277,7 @@ class branch_and_bound_t { dual::status_t solve_node_lp(mip_node_t* node_ptr, lp_problem_t& leaf_problem, lp_solution_t& leaf_solution, + std::vector& leaf_edge_norms, basis_update_mpf_t& basis_factors, std::vector& basic_list, std::vector& nonbasic_list, @@ -269,6 +296,7 @@ class branch_and_bound_t { search_tree_t& search_tree, lp_problem_t& leaf_problem, lp_solution_t& leaf_solution, + std::vector& leaf_edge_norms, bnb_worker_type_t thread_type, dual::status_t lp_status, logger_t& log); diff --git a/cpp/src/dual_simplex/crossover.cpp b/cpp/src/dual_simplex/crossover.cpp index 8ee3fb0ce4..81d5ec1e6d 100644 --- a/cpp/src/dual_simplex/crossover.cpp +++ b/cpp/src/dual_simplex/crossover.cpp @@ -25,7 +25,7 @@ crossover_status_t return_to_status(int status) { if (status == -1) { return crossover_status_t::TIME_LIMIT; - } else if (status == -2) { + } else if (status == CONCURRENT_HALT_RETURN) { return crossover_status_t::CONCURRENT_LIMIT; } else { return crossover_status_t::NUMERICAL_ISSUES; @@ -502,7 +502,32 @@ i_t dual_push(const lp_problem_t& lp, std::vector q(m); std::vector deficient; std::vector slacks_needed; - factorize_basis(lp.A, settings, basic_list, L, U, p, pinv, q, deficient, slacks_needed); + i_t rank = + factorize_basis(lp.A, settings, basic_list, L, U, p, pinv, q, deficient, slacks_needed); + if (rank == CONCURRENT_HALT_RETURN) { + return CONCURRENT_HALT_RETURN; + } else if (rank != m) { + settings.log.printf("Failed to factorize basis. rank %d m %d\n", rank, m); + basis_repair(lp.A, + settings, + lp.lower, + lp.upper, + deficient, + slacks_needed, + basic_list, + nonbasic_list, + vstatus); + rank = + factorize_basis(lp.A, settings, basic_list, L, U, p, pinv, q, deficient, slacks_needed); + if (rank == CONCURRENT_HALT_RETURN) { + return CONCURRENT_HALT_RETURN; + } else if (rank == -1) { + settings.log.printf("Failed to factorize basis after repair. rank %d m %d\n", rank, m); + return -1; + } else { + settings.log.printf("Basis repaired\n"); + } + } reorder_basic_list(q, basic_list); // Reordering the basic list causes us to mess up the superbasic list index // so we need to update it @@ -535,7 +560,7 @@ i_t dual_push(const lp_problem_t& lp, } if (settings.concurrent_halt != nullptr && *settings.concurrent_halt == 1) { settings.log.printf("Concurrent halt\n"); - return -2; + return CONCURRENT_HALT_RETURN; } } @@ -783,7 +808,9 @@ i_t primal_push(const lp_problem_t& lp, std::vector slacks_needed; i_t rank = factorize_basis(lp.A, settings, basic_list, L, U, p, pinv, q, deficient, slacks_needed); - if (rank != m) { + if (rank == CONCURRENT_HALT_RETURN) { + return CONCURRENT_HALT_RETURN; + } else if (rank != m) { settings.log.debug("Failed to factorize basis. rank %d m %d\n", rank, m); basis_repair(lp.A, settings, @@ -794,8 +821,11 @@ i_t primal_push(const lp_problem_t& lp, basic_list, nonbasic_list, vstatus); - if (factorize_basis( - lp.A, settings, basic_list, L, U, p, pinv, q, deficient, slacks_needed) == -1) { + rank = + factorize_basis(lp.A, settings, basic_list, L, U, p, pinv, q, deficient, slacks_needed); + if (rank == CONCURRENT_HALT_RETURN) { + return CONCURRENT_HALT_RETURN; + } else if (rank == -1) { settings.log.printf("Failed to factorize basis after repair. rank %d m %d\n", rank, m); return -1; } else { @@ -833,7 +863,7 @@ i_t primal_push(const lp_problem_t& lp, } if (settings.concurrent_halt != nullptr && *settings.concurrent_halt == 1) { settings.log.printf("Concurrent halt\n"); - return -2; + return CONCURRENT_HALT_RETURN; } } @@ -1137,6 +1167,7 @@ crossover_status_t crossover(const lp_problem_t& lp, std::vector slacks_needed; rank = factorize_basis(lp.A, settings, basic_list, L, U, p, pinv, q, deficient, slacks_needed); + if (rank == CONCURRENT_HALT_RETURN) { return crossover_status_t::CONCURRENT_LIMIT; } if (rank != m) { settings.log.debug("Failed to factorize basis. rank %d m %d\n", rank, m); basis_repair(lp.A, @@ -1148,8 +1179,10 @@ crossover_status_t crossover(const lp_problem_t& lp, basic_list, nonbasic_list, vstatus); - if (factorize_basis(lp.A, settings, basic_list, L, U, p, pinv, q, deficient, slacks_needed) == - -1) { + rank = factorize_basis(lp.A, settings, basic_list, L, U, p, pinv, q, deficient, slacks_needed); + if (rank == CONCURRENT_HALT_RETURN) { + return crossover_status_t::CONCURRENT_LIMIT; + } else if (rank == -1) { settings.log.printf("Failed to factorize basis after repair. rank %d m %d\n", rank, m); return crossover_status_t::NUMERICAL_ISSUES; } else { @@ -1218,8 +1251,8 @@ crossover_status_t crossover(const lp_problem_t& lp, i_t primal_push_status = primal_push( lp, settings, start_time, solution, ft, basic_list, nonbasic_list, superbasic_list, vstatus); if (primal_push_status < 0) { return return_to_status(primal_push_status); } - print_crossover_info(lp, settings, vstatus, solution, "Primal push complete"); compute_dual_solution_from_basis(lp, ft, basic_list, nonbasic_list, solution.y, solution.z); + print_crossover_info(lp, settings, vstatus, solution, "Primal push complete"); } else { settings.log.printf("No primal push needed. No superbasic variables\n"); } @@ -1336,7 +1369,9 @@ crossover_status_t crossover(const lp_problem_t& lp, get_basis_from_vstatus(m, vstatus, basic_list, nonbasic_list, superbasic_list); rank = factorize_basis(lp.A, settings, basic_list, L, U, p, pinv, q, deficient, slacks_needed); - if (rank != m) { + if (rank == CONCURRENT_HALT_RETURN) { + return crossover_status_t::CONCURRENT_LIMIT; + } else if (rank != m) { settings.log.debug("Failed to factorize basis. rank %d m %d\n", rank, m); basis_repair(lp.A, settings, @@ -1347,8 +1382,11 @@ crossover_status_t crossover(const lp_problem_t& lp, basic_list, nonbasic_list, vstatus); - if (factorize_basis( - lp.A, settings, basic_list, L, U, p, pinv, q, deficient, slacks_needed) == -1) { + rank = + factorize_basis(lp.A, settings, basic_list, L, U, p, pinv, q, deficient, slacks_needed); + if (rank == CONCURRENT_HALT_RETURN) { + return crossover_status_t::CONCURRENT_LIMIT; + } else if (rank == -1) { settings.log.printf("Failed to factorize basis after repair. rank %d m %d\n", rank, m); return crossover_status_t::NUMERICAL_ISSUES; } else { @@ -1358,8 +1396,7 @@ crossover_status_t crossover(const lp_problem_t& lp, reorder_basic_list(q, basic_list); ft.reset(L, U, p); - compute_dual_solution_from_basis(lp, ft, basic_list, nonbasic_list, solution.y, solution.z); - + solution = phase1_solution; i_t num_flips = 0; for (i_t j = 0; j < n; ++j) { if (vstatus[j] == variable_status_t::BASIC) { continue; } @@ -1376,20 +1413,23 @@ crossover_status_t crossover(const lp_problem_t& lp, } } settings.log.debug("Num flips %d\n", num_flips); - solution = phase1_solution; print_crossover_info(lp, settings, vstatus, solution, "Dual phase 1 complete"); - std::vector edge_norms; - dual::status_t status = dual_phase2( - 2, iter == 0 ? 1 : 0, start_time, lp, settings, vstatus, solution, iter, edge_norms); - if (toc(start_time) > settings.time_limit) { - settings.log.printf("Time limit exceeded\n"); - return crossover_status_t::TIME_LIMIT; - } - if (settings.concurrent_halt != nullptr && *settings.concurrent_halt == 1) { - settings.log.printf("Concurrent halt\n"); - return crossover_status_t::CONCURRENT_LIMIT; + dual_infeas = dual_infeasibility(lp, settings, vstatus, solution.z); + dual::status_t status = dual::status_t::NUMERICAL; + if (dual_infeas <= settings.dual_tol) { + std::vector edge_norms; + status = dual_phase2( + 2, iter == 0 ? 1 : 0, start_time, lp, settings, vstatus, solution, iter, edge_norms); + if (toc(start_time) > settings.time_limit) { + settings.log.printf("Time limit exceeded\n"); + return crossover_status_t::TIME_LIMIT; + } + if (settings.concurrent_halt != nullptr && *settings.concurrent_halt == 1) { + settings.log.printf("Concurrent halt\n"); + return crossover_status_t::CONCURRENT_LIMIT; + } + solution.iterations += iter; } - solution.iterations += iter; primal_infeas = primal_infeasibility(lp, settings, vstatus, solution.x); dual_infeas = dual_infeasibility(lp, settings, vstatus, solution.z); primal_res = primal_residual(lp, solution); diff --git a/cpp/src/dual_simplex/cuts.cpp b/cpp/src/dual_simplex/cuts.cpp new file mode 100644 index 0000000000..be3f3001d8 --- /dev/null +++ b/cpp/src/dual_simplex/cuts.cpp @@ -0,0 +1,2822 @@ +/* clang-format off */ +/* + * SPDX-FileCopyrightText: Copyright (c) 2025-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. + * SPDX-License-Identifier: Apache-2.0 + */ +/* clang-format on */ + +#include +#include +#include +#include + +namespace cuopt::linear_programming::dual_simplex { + +template +void cut_pool_t::add_cut(cut_type_t cut_type, + const sparse_vector_t& cut, + f_t rhs) +{ + // TODO: Need to deduplicate cuts and only add if the cut is not already in the pool + + for (i_t p = 0; p < cut.i.size(); p++) { + const i_t j = cut.i[p]; + if (j >= original_vars_) { + settings_.log.printf( + "Cut has variable %d that is greater than original_vars_ %d\n", j, original_vars_); + return; + } + } + + sparse_vector_t cut_squeezed; + cut.squeeze(cut_squeezed); + if (cut_squeezed.i.size() == 0) { + settings_.log.printf("Cut has no coefficients\n"); + return; + } + cut_storage_.append_row(cut_squeezed); + rhs_storage_.push_back(rhs); + cut_type_.push_back(cut_type); + cut_age_.push_back(0); +} + +template +f_t cut_pool_t::cut_distance(i_t row, + const std::vector& x, + f_t& cut_violation, + f_t& cut_norm) +{ + const i_t row_start = cut_storage_.row_start[row]; + const i_t row_end = cut_storage_.row_start[row + 1]; + f_t cut_x = 0.0; + f_t dot = 0.0; + for (i_t p = row_start; p < row_end; p++) { + const i_t j = cut_storage_.j[p]; + const f_t cut_coeff = cut_storage_.x[p]; + cut_x += cut_coeff * x[j]; + dot += cut_coeff * cut_coeff; + } + cut_violation = rhs_storage_[row] - cut_x; + cut_norm = std::sqrt(dot); + const f_t distance = cut_violation / cut_norm; + return distance; +} + +template +f_t cut_pool_t::cut_density(i_t row) +{ + const i_t row_start = cut_storage_.row_start[row]; + const i_t row_end = cut_storage_.row_start[row + 1]; + const i_t cut_nz = row_end - row_start; + const i_t original_vars = original_vars_; + return static_cast(cut_nz) / original_vars; +} + +template +f_t cut_pool_t::cut_orthogonality(i_t i, i_t j) +{ + const i_t i_start = cut_storage_.row_start[i]; + const i_t i_end = cut_storage_.row_start[i + 1]; + const i_t i_nz = i_end - i_start; + const i_t j_start = cut_storage_.row_start[j]; + const i_t j_end = cut_storage_.row_start[j + 1]; + const i_t j_nz = j_end - j_start; + + f_t dot = sparse_dot(cut_storage_.j.data() + i_start, + cut_storage_.x.data() + i_start, + i_nz, + cut_storage_.j.data() + j_start, + cut_storage_.x.data() + j_start, + j_nz); + + f_t norm_i = cut_norms_[i]; + f_t norm_j = cut_norms_[j]; + return 1.0 - std::abs(dot) / (norm_i * norm_j); +} + +template +void cut_pool_t::score_cuts(std::vector& x_relax) +{ + const f_t min_cut_distance = 1e-4; + cut_distances_.resize(cut_storage_.m, 0.0); + cut_norms_.resize(cut_storage_.m, 0.0); + + const bool verbose = false; + for (i_t i = 0; i < cut_storage_.m; i++) { + f_t violation; + f_t cut_dist = cut_distance(i, x_relax, violation, cut_norms_[i]); + cut_distances_[i] = cut_dist <= min_cut_distance ? 0.0 : cut_dist; + if (verbose) { + settings_.log.printf("Cut %d type %d distance %+e violation %+e cut_norm %e\n", + i, + static_cast(cut_type_[i]), + cut_distances_[i], + violation, + cut_norms_[i]); + } + } + + std::vector sorted_indices; + best_score_last_permutation(cut_distances_, sorted_indices); + + const i_t max_cuts = 2000; + const f_t min_orthogonality = settings_.cut_min_orthogonality; + best_cuts_.reserve(std::min(max_cuts, cut_storage_.m)); + best_cuts_.clear(); + scored_cuts_ = 0; + + if (!sorted_indices.empty()) { + const i_t i = sorted_indices.back(); + sorted_indices.pop_back(); + best_cuts_.push_back(i); + scored_cuts_++; + } + + while (scored_cuts_ < max_cuts && !sorted_indices.empty()) { + const i_t i = sorted_indices.back(); + sorted_indices.pop_back(); + + if (cut_distances_[i] <= min_cut_distance) { break; } + + f_t cut_ortho = 1.0; + const i_t best_cuts_size = best_cuts_.size(); + for (i_t k = 0; k < best_cuts_size; k++) { + const i_t j = best_cuts_[k]; + cut_ortho = std::min(cut_ortho, cut_orthogonality(i, j)); + } + if (cut_ortho >= min_orthogonality) { + best_cuts_.push_back(i); + scored_cuts_++; + } + } +} + +template +i_t cut_pool_t::get_best_cuts(csr_matrix_t& best_cuts, + std::vector& best_rhs, + std::vector& best_cut_types) +{ + best_cuts.m = 0; + best_cuts.n = original_vars_; + best_cuts.row_start.clear(); + best_cuts.j.clear(); + best_cuts.x.clear(); + best_cuts.row_start.reserve(scored_cuts_ + 1); + best_cuts.row_start.push_back(0); + best_rhs.clear(); + best_rhs.reserve(scored_cuts_); + best_cut_types.clear(); + best_cut_types.reserve(scored_cuts_); + + for (i_t i : best_cuts_) { + sparse_vector_t cut(cut_storage_, i); + cut.negate(); + best_cuts.append_row(cut); + best_rhs.push_back(-rhs_storage_[i]); + best_cut_types.push_back(cut_type_[i]); + } + + age_cuts(); + + return static_cast(best_cuts_.size()); +} + +template +void cut_pool_t::age_cuts() +{ + for (i_t i = 0; i < cut_age_.size(); i++) { + cut_age_[i]++; + } +} + +template +void cut_pool_t::drop_cuts() +{ + // TODO: Implement this +} + +template +knapsack_generation_t::knapsack_generation_t( + const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + csr_matrix_t& Arow, + const std::vector& new_slacks, + const std::vector& var_types) + : settings_(settings) +{ + const bool verbose = false; + knapsack_constraints_.reserve(lp.num_rows); + + is_slack_.resize(lp.num_cols, 0); + for (i_t j : new_slacks) { + is_slack_[j] = 1; + } + + for (i_t i = 0; i < lp.num_rows; i++) { + const i_t row_start = Arow.row_start[i]; + const i_t row_end = Arow.row_start[i + 1]; + const i_t row_len = row_end - row_start; + if (row_len < 3) { continue; } + bool is_knapsack = true; + f_t sum_pos = 0.0; + for (i_t p = row_start; p < row_end; p++) { + const i_t j = Arow.j[p]; + if (is_slack_[j]) { continue; } + const f_t aj = Arow.x[p]; + if (std::abs(aj - std::round(aj)) > settings.integer_tol) { + is_knapsack = false; + break; + } + if (var_types[j] != variable_type_t::INTEGER || lp.lower[j] != 0.0 || lp.upper[j] != 1.0) { + is_knapsack = false; + break; + } + if (aj < 0.0) { + is_knapsack = false; + break; + } + sum_pos += aj; + } + + if (is_knapsack) { + const f_t beta = lp.rhs[i]; + if (std::abs(beta - std::round(beta)) <= settings.integer_tol) { + if (beta > 0.0 && beta <= sum_pos && std::abs(sum_pos / (row_len - 1) - beta) > 1e-3) { + if (verbose) { + settings.log.printf( + "Knapsack constraint %d row len %d beta %e sum_pos %e sum_pos / (row_len - 1) %e\n", + i, + row_len, + beta, + sum_pos, + sum_pos / (row_len - 1)); + } + knapsack_constraints_.push_back(i); + } + } + } + } + +#ifdef PRINT_KNAPSACK_INFO + i_t num_knapsack_constraints = knapsack_constraints_.size(); + settings.log.printf("Number of knapsack constraints %d\n", num_knapsack_constraints); +#endif +} + +template +i_t knapsack_generation_t::generate_knapsack_cuts( + const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + csr_matrix_t& Arow, + const std::vector& new_slacks, + const std::vector& var_types, + const std::vector& xstar, + i_t knapsack_row, + sparse_vector_t& cut, + f_t& cut_rhs) +{ + const bool verbose = false; + // Get the row associated with the knapsack constraint + sparse_vector_t knapsack_inequality(Arow, knapsack_row); + f_t knapsack_rhs = lp.rhs[knapsack_row]; + + // Remove the slacks from the inequality + f_t seperation_rhs = 0.0; + if (verbose) { settings.log.printf(" Knapsack : "); } + for (i_t k = 0; k < knapsack_inequality.i.size(); k++) { + const i_t j = knapsack_inequality.i[k]; + if (is_slack_[j]) { + knapsack_inequality.x[k] = 0.0; + } else { + if (verbose) { settings.log.printf(" %g x%d +", knapsack_inequality.x[k], j); } + seperation_rhs += knapsack_inequality.x[k]; + } + } + if (verbose) { settings.log.printf(" <= %g\n", knapsack_rhs); } + seperation_rhs -= (knapsack_rhs + 1); + + if (verbose) { + settings.log.printf("\t"); + for (i_t k = 0; k < knapsack_inequality.i.size(); k++) { + const i_t j = knapsack_inequality.i[k]; + if (!is_slack_[j]) { + if (std::abs(xstar[j]) > 1e-3) { settings.log.printf("x_relax[%d]= %g ", j, xstar[j]); } + } + } + settings.log.printf("\n"); + + settings.log.printf("seperation_rhs %g\n", seperation_rhs); + } + if (seperation_rhs <= 0.0) { return -1; } + + std::vector values; + values.resize(knapsack_inequality.i.size() - 1); + std::vector weights; + weights.resize(knapsack_inequality.i.size() - 1); + i_t h = 0; + f_t objective_constant = 0.0; + for (i_t k = 0; k < knapsack_inequality.i.size(); k++) { + const i_t j = knapsack_inequality.i[k]; + if (!is_slack_[j]) { + const f_t vj = std::min(1.0, std::max(0.0, 1.0 - xstar[j])); + objective_constant += vj; + values[h] = vj; + weights[h] = knapsack_inequality.x[k]; + h++; + } + } + std::vector solution; + solution.resize(knapsack_inequality.i.size() - 1); + + if (verbose) { settings.log.printf("Calling solve_knapsack_problem\n"); } + f_t objective = solve_knapsack_problem(values, weights, seperation_rhs, solution); + if (std::isnan(objective)) { return -1; } + if (verbose) { + settings.log.printf("objective %e objective_constant %e\n", objective, objective_constant); + } + f_t seperation_value = -objective + objective_constant; + if (verbose) { settings.log.printf("seperation_value %e\n", seperation_value); } + const f_t tol = 1e-6; + if (seperation_value >= 1.0 - tol) { return -1; } + + i_t cover_size = 0; + for (i_t k = 0; k < solution.size(); k++) { + if (solution[k] == 0.0) { cover_size++; } + } + + cut.i.clear(); + cut.x.clear(); + cut.i.reserve(cover_size); + cut.x.reserve(cover_size); + + h = 0; + for (i_t k = 0; k < knapsack_inequality.i.size(); k++) { + const i_t j = knapsack_inequality.i[k]; + if (!is_slack_[j]) { + if (solution[h] == 0.0) { + cut.i.push_back(j); + cut.x.push_back(-1.0); + } + h++; + } + } + cut_rhs = -cover_size + 1; + cut.sort(); + + // The cut is in the form: - sum_{j in cover} x_j >= -cover_size + 1 + // Which is equivalent to: sum_{j in cover} x_j <= cover_size - 1 + + // Verify the cut is violated + f_t dot = cut.dot(xstar); + f_t violation = dot - cut_rhs; + if (verbose) { + settings.log.printf("Knapsack cut %d violation %e < 0\n", knapsack_row, violation); + } + + if (violation >= -tol) { return -1; } + +#ifdef PRINT_KNAPSACK_CUT + settings.log.printf("knapsack cut (cover %d): \n", cover_size); + for (i_t k = 0; k < cut.i.size(); k++) { + settings.log.printf("x%d coeff %g value %g\n", cut.i[k], -cut.x[k], xstar[cut.i[k]]); + } + settings.log.printf("cut_rhs %g\n", -cut_rhs); +#endif + return 0; +} + +template +f_t knapsack_generation_t::greedy_knapsack_problem(const std::vector& values, + const std::vector& weights, + f_t rhs, + std::vector& solution) +{ + i_t n = weights.size(); + solution.assign(n, 0.0); + + // Build permutation + std::vector perm(n); + std::iota(perm.begin(), perm.end(), 0); + + std::vector ratios; + ratios.resize(n); + for (i_t i = 0; i < n; i++) { + ratios[i] = values[i] / weights[i]; + } + + // Sort by value / weight ratio + std::sort(perm.begin(), perm.end(), [&](i_t i, i_t j) { return ratios[i] > ratios[j]; }); + + // Greedy select items with the best value / weight ratio until the remaining capacity is + // exhausted + f_t remaining = rhs; + f_t total_value = 0.0; + + for (i_t j : perm) { + if (weights[j] <= remaining) { + solution[j] = 1.0; + remaining -= weights[j]; + total_value += values[j]; + } + } + + // Best single-item fallback + f_t best_single_value = 0.0; + i_t best_single_idx = -1; + + for (i_t j = 0; j < n; ++j) { + if (weights[j] <= rhs && values[j] > best_single_value) { + best_single_value = values[j]; + best_single_idx = j; + } + } + + if (best_single_value > total_value) { + solution.assign(n, 0.0); + solution[best_single_idx] = 1.0; + return best_single_value; + } + + return total_value; +} + +template +f_t knapsack_generation_t::solve_knapsack_problem(const std::vector& values, + const std::vector& weights, + f_t rhs, + std::vector& solution) +{ + // Solve the knapsack problem + // maximize sum_{j=0}^n values[j] * solution[j] + // subject to sum_{j=0}^n weights[j] * solution[j] <= rhs + // values: values of the items + // weights: weights of the items + // return the value of the solution + + // Using approximate dynamic programming + + i_t n = weights.size(); + f_t objective = std::numeric_limits::quiet_NaN(); + + // Compute the maximum value + f_t vmax = *std::max_element(values.begin(), values.end()); + + // Check if all the values are integers + bool all_integers = true; + const f_t integer_tol = 1e-5; + for (i_t j = 0; j < n; j++) { + if (std::abs(values[j] - std::round(values[j])) > integer_tol) { + all_integers = false; + break; + } + } + + const bool verbose = false; + + if (verbose) { settings_.log.printf("all_integers %d\n", all_integers); } + + // Compute the scaling factor and comptue the scaled integer values + f_t scale = 1.0; + std::vector scaled_values(n); + if (all_integers) { + for (i_t j = 0; j < n; j++) { + scaled_values[j] = static_cast(std::floor(values[j])); + } + } else { + const f_t epsilon = 0.1; + scale = epsilon * vmax / static_cast(n); + if (scale <= 0.0) { return std::numeric_limits::quiet_NaN(); } + if (verbose) { + settings_.log.printf("scale %g epsilon %g vmax %g n %d\n", scale, epsilon, vmax, n); + } + for (i_t i = 0; i < n; ++i) { + scaled_values[i] = static_cast(std::floor(values[i] / scale)); + } + } + + i_t sum_value = std::accumulate(scaled_values.begin(), scaled_values.end(), 0); + const i_t INT_INF = std::numeric_limits::max() / 2; + if (verbose) { settings_.log.printf("sum value %d\n", sum_value); } + const i_t max_size = 10000; + if (sum_value <= 0.0 || sum_value >= max_size) { + if (verbose) { + settings_.log.printf("sum value %d is negative or too large using greedy solution\n", + sum_value); + } + return greedy_knapsack_problem(values, weights, rhs, solution); + } + + // dp(j, v) = minimum weight using first j items to get value v + dense_matrix_t dp(n + 1, sum_value + 1, INT_INF); + dense_matrix_t take(n + 1, sum_value + 1, 0); + dp(0, 0) = 0; + + // 4. Dynamic programming + for (i_t j = 1; j <= n; ++j) { + for (i_t v = 0; v <= sum_value; ++v) { + // Do not take item i-1 + dp(j, v) = dp(j - 1, v); + + // Take item j-1 if possible + if (v >= scaled_values[j - 1]) { + i_t candidate = + dp(j - 1, v - scaled_values[j - 1]) + static_cast(std::floor(weights[j - 1])); + if (candidate < dp(j, v)) { + dp(j, v) = candidate; + take(j, v) = 1; + } + } + } + } + + // 5. Find best achievable value within capacity + i_t best_value = 0; + for (i_t v = 0; v <= sum_value; ++v) { + if (dp(n, v) <= rhs) { best_value = v; } + } + + // 6. Backtrack to recover solution + i_t v = best_value; + for (i_t j = n; j >= 1; --j) { + if (take(j, v)) { + solution[j - 1] = 1.0; + v -= scaled_values[j - 1]; + } else { + solution[j - 1] = 0.0; + } + } + + objective = best_value * scale; + return objective; +} + +template +void cut_generation_t::generate_cuts(const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + csr_matrix_t& Arow, + const std::vector& new_slacks, + const std::vector& var_types, + basis_update_mpf_t& basis_update, + const std::vector& xstar, + const std::vector& basic_list, + const std::vector& nonbasic_list) +{ + // Generate Gomory and CG Cuts + if (settings.mixed_integer_gomory_cuts != 0 || settings.strong_chvatal_gomory_cuts != 0) { + f_t cut_start_time = tic(); + generate_gomory_cuts( + lp, settings, Arow, new_slacks, var_types, basis_update, xstar, basic_list, nonbasic_list); + f_t cut_generation_time = toc(cut_start_time); + if (cut_generation_time > 1.0) { + settings.log.debug("Gomory and CG cut generation time %.2f seconds\n", cut_generation_time); + } + } + + // Generate Knapsack cuts + if (settings.knapsack_cuts != 0) { + f_t cut_start_time = tic(); + generate_knapsack_cuts(lp, settings, Arow, new_slacks, var_types, xstar); + f_t cut_generation_time = toc(cut_start_time); + if (cut_generation_time > 1.0) { + settings.log.debug("Knapsack cut generation time %.2f seconds\n", cut_generation_time); + } + } + + // Generate MIR and CG cuts + if (settings.mir_cuts != 0 || settings.strong_chvatal_gomory_cuts != 0) { + f_t cut_start_time = tic(); + generate_mir_cuts(lp, settings, Arow, new_slacks, var_types, xstar); + f_t cut_generation_time = toc(cut_start_time); + if (cut_generation_time > 1.0) { + settings.log.debug("MIR and CG cut generation time %.2f seconds\n", cut_generation_time); + } + } +} + +template +void cut_generation_t::generate_knapsack_cuts( + const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + csr_matrix_t& Arow, + const std::vector& new_slacks, + const std::vector& var_types, + const std::vector& xstar) +{ + if (knapsack_generation_.num_knapsack_constraints() > 0) { + for (i_t knapsack_row : knapsack_generation_.get_knapsack_constraints()) { + sparse_vector_t cut(lp.num_cols, 0); + f_t cut_rhs; + i_t knapsack_status = knapsack_generation_.generate_knapsack_cuts( + lp, settings, Arow, new_slacks, var_types, xstar, knapsack_row, cut, cut_rhs); + if (knapsack_status == 0) { cut_pool_.add_cut(cut_type_t::KNAPSACK, cut, cut_rhs); } + } + } +} + +template +void cut_generation_t::generate_mir_cuts( + const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + csr_matrix_t& Arow, + const std::vector& new_slacks, + const std::vector& var_types, + const std::vector& xstar) +{ + f_t mir_start_time = tic(); + mixed_integer_rounding_cut_t mir(lp, settings, new_slacks, xstar); + strong_cg_cut_t cg(lp, var_types, xstar); + + std::vector slack_map(lp.num_rows, -1); + for (i_t slack : new_slacks) { + const i_t col_start = lp.A.col_start[slack]; + const i_t col_end = lp.A.col_start[slack + 1]; + const i_t col_len = col_end - col_start; + assert(col_len == 1); + const i_t i = lp.A.i[col_start]; + slack_map[i] = slack; + } + + // Compute initial scores for all rows + std::vector score(lp.num_rows, 0.0); + for (i_t i = 0; i < lp.num_rows; i++) { + const i_t row_start = Arow.row_start[i]; + const i_t row_end = Arow.row_start[i + 1]; + + const i_t row_nz = row_end - row_start; + i_t num_integer_in_row = 0; + i_t num_continuous_in_row = 0; + for (i_t p = row_start; p < row_end; p++) { + const i_t j = Arow.j[p]; + if (var_types[j] == variable_type_t::INTEGER) { + num_integer_in_row++; + } else { + num_continuous_in_row++; + } + } + + if (num_integer_in_row == 0) { + score[i] = 0.0; + + } else { + f_t nz_score = lp.num_cols - row_nz; + + const i_t slack = slack_map[i]; + assert(slack >= 0); + const f_t slack_value = xstar[slack]; + + f_t slack_score = -std::log10(1e-16 + std::abs(slack_value)); + + const f_t nz_weight = 1.0; + const f_t slack_weight = 1.0; + const f_t integer_weight = 1.0; + + score[i] = + nz_weight * nz_score + slack_weight * slack_score + integer_weight * num_integer_in_row; + } + } + + // Sort the rows by score + std::vector sorted_indices; + best_score_last_permutation(score, sorted_indices); + + // These data structures are used to track the rows that have been aggregated + // The invariant is that aggregated_rows is empty and aggregated_mark is all zeros + // at the beginning of each iteration of the for loop below + std::vector aggregated_rows; + std::vector aggregated_mark(lp.num_rows, 0); + + const i_t max_cuts = std::min(lp.num_rows, 1000); + f_t work_estimate = 0.0; + for (i_t h = 0; h < max_cuts; h++) { + // Get the row with the highest score + const i_t i = sorted_indices.back(); + sorted_indices.pop_back(); + const f_t max_score = score[i]; + + const i_t row_nz = Arow.row_start[i + 1] - Arow.row_start[i]; + const i_t slack = slack_map[i]; + const f_t slack_value = xstar[slack]; + + if (max_score <= 0.0) { break; } + if (work_estimate > 2e9) { break; } + + sparse_vector_t inequality(Arow, i); + work_estimate += inequality.i.size(); + f_t inequality_rhs = lp.rhs[i]; + const bool generate_cg_cut = settings.strong_chvatal_gomory_cuts != 0; + f_t fractional_part_rhs = fractional_part(inequality_rhs); + if (generate_cg_cut && fractional_part_rhs > 1e-6 && fractional_part_rhs < (1 - 1e-6)) { + // Try to generate a CG cut + sparse_vector_t cg_inequality = inequality; + f_t cg_inequality_rhs = inequality_rhs; + if (fractional_part(inequality_rhs) < 0.5) { + // Multiply by -1 to force the fractional part to be greater than 0.5 + cg_inequality_rhs *= -1; + cg_inequality.negate(); + } + sparse_vector_t cg_cut(lp.num_cols, 0); + f_t cg_cut_rhs; + i_t cg_status = cg.generate_strong_cg_cut( + lp, settings, var_types, cg_inequality, cg_inequality_rhs, xstar, cg_cut, cg_cut_rhs); + if (cg_status == 0) { cut_pool_.add_cut(cut_type_t::CHVATAL_GOMORY, cg_cut, cg_cut_rhs); } + } + + // Remove the slack from the equality to get an inequality + work_estimate += inequality.i.size(); + i_t negate_inequality = 1; + for (i_t k = 0; k < inequality.i.size(); k++) { + const i_t j = inequality.i[k]; + if (j == slack) { + if (inequality.x[k] != 1.0) { + if (inequality.x[k] == -1.0 && lp.lower[j] >= 0.0) { + negate_inequality = 0; + } else { + settings.log.debug("Bad slack %d in inequality: aj %e lo %e up %e\n", + j, + inequality.x[k], + lp.lower[j], + lp.upper[j]); + negate_inequality = -1; + break; + } + } + inequality.x[k] = 0.0; + } + } + + if (negate_inequality == -1) { continue; } + + if (negate_inequality) { + // inequaility'*x <= inequality_rhs + // But for MIR we need: inequality'*x >= inequality_rhs + inequality_rhs *= -1; + inequality.negate(); + work_estimate += inequality.i.size(); + } + // We should now have: inequality'*x >= inequality_rhs + + // Transform the relaxation solution + std::vector transformed_xstar; + mir.relaxation_to_nonnegative(lp, xstar, transformed_xstar); + work_estimate += transformed_xstar.size(); + + sparse_vector_t cut(lp.num_cols, 0); + f_t cut_rhs; + bool add_cut = false; + i_t num_aggregated = 0; + const i_t max_aggregated = 6; + work_estimate += lp.num_cols; + + while (!add_cut && num_aggregated < max_aggregated) { + sparse_vector_t transformed_inequality; + inequality.squeeze(transformed_inequality); + f_t transformed_rhs = inequality_rhs; + work_estimate += transformed_inequality.i.size(); + + mir.to_nonnegative(lp, transformed_inequality, transformed_rhs); + work_estimate += transformed_inequality.i.size(); + std::vector> transformed_cuts; + std::vector transformed_cut_rhs; + std::vector transformed_violations; + + // Generate cut for delta = 1 + { + sparse_vector_t cut_1(lp.num_cols, 0); + f_t cut_1_rhs; + mir.generate_cut_nonnegative( + transformed_inequality, transformed_rhs, var_types, cut_1, cut_1_rhs); + f_t cut_1_violation = mir.compute_violation(cut_1, cut_1_rhs, transformed_xstar); + if (cut_1_violation > 1e-6) { + transformed_cuts.push_back(cut_1); + transformed_cut_rhs.push_back(cut_1_rhs); + transformed_violations.push_back(cut_1_violation); + } + work_estimate += transformed_inequality.i.size(); + } + + // Generate a cut for delta = max { |a_j|, j in I} + { + f_t max_coeff = 0.0; + for (i_t k = 0; k < transformed_inequality.i.size(); k++) { + const i_t j = transformed_inequality.i[k]; + if (var_types[j] == variable_type_t::INTEGER) { + const f_t abs_aj = std::abs(transformed_inequality.x[k]); + if (abs_aj > max_coeff) { max_coeff = abs_aj; } + } + } + work_estimate += transformed_inequality.i.size(); + + if (max_coeff > 1e-6 && max_coeff != 1.0) { + sparse_vector_t scaled_inequality = transformed_inequality; + const i_t nz = transformed_inequality.i.size(); + for (i_t k = 0; k < nz; k++) { + scaled_inequality.x[k] /= max_coeff; + } + const f_t scaled_rhs = transformed_rhs / max_coeff; + sparse_vector_t cut_2(lp.num_cols, 0); + f_t cut_2_rhs; + mir.generate_cut_nonnegative(scaled_inequality, scaled_rhs, var_types, cut_2, cut_2_rhs); + f_t cut_2_violation = mir.compute_violation(cut_2, cut_2_rhs, transformed_xstar); + if (cut_2_violation > 1e-6) { + transformed_cuts.push_back(cut_2); + transformed_cut_rhs.push_back(cut_2_rhs); + transformed_violations.push_back(cut_2_violation); + } + work_estimate += 5 * transformed_inequality.i.size(); + } + } + + if (!transformed_violations.empty()) { + std::vector permuted(transformed_violations.size()); + std::iota(permuted.begin(), permuted.end(), 0); + std::sort(permuted.begin(), permuted.end(), [&](i_t i, i_t j) { + return transformed_violations[i] > transformed_violations[j]; + }); + work_estimate += transformed_violations.size() * std::log2(transformed_violations.size()); + // Get the biggest violation + const i_t best_index = permuted[0]; + f_t max_viol = transformed_violations[best_index]; + cut = transformed_cuts[best_index]; + cut_rhs = transformed_cut_rhs[best_index]; + + if (max_viol > 1e-6) { + // TODO: Divide by 1/2*violation, 1/4*violation, 1/8*violation + // Transform back to the original variables + mir.to_original(lp, cut, cut_rhs); + mir.remove_small_coefficients(lp.lower, lp.upper, cut, cut_rhs); + mir.substitute_slacks(lp, Arow, cut, cut_rhs); + f_t viol = mir.compute_violation(cut, cut_rhs, xstar); + work_estimate += 10 * cut.i.size(); + add_cut = true; + } + } + + if (add_cut) { + if (settings.mir_cuts != 0) { + cut_pool_.add_cut(cut_type_t::MIXED_INTEGER_ROUNDING, cut, cut_rhs); + } + break; + } else { + // Perform aggregation to try and find a cut + + // Find all the continuous variables in the inequality + i_t num_continuous = 0; + f_t max_off_bound = 0.0; + i_t max_off_bound_var = -1; + for (i_t p = 0; p < inequality.i.size(); p++) { + const i_t j = inequality.i[p]; + if (var_types[j] == variable_type_t::CONTINUOUS) { + num_continuous++; + + const f_t off_lower = lp.lower[j] > -inf ? xstar[j] - lp.lower[j] : std::abs(xstar[j]); + const f_t off_upper = lp.upper[j] < inf ? lp.upper[j] - xstar[j] : std::abs(xstar[j]); + const f_t off_bound = std::max(off_lower, off_upper); + const i_t col_start = lp.A.col_start[j]; + const i_t col_end = lp.A.col_start[j + 1]; + const i_t col_len = col_end - col_start; + if (off_bound > max_off_bound && col_len > 1) { + max_off_bound = off_bound; + max_off_bound_var = j; + } + } + } + work_estimate += 10 * inequality.i.size(); + + if (num_continuous == 0 || max_off_bound < 1e-6) { break; } + + // The variable that is farthest from its bound is used as a pivot + if (max_off_bound_var >= 0) { + const i_t col_start = lp.A.col_start[max_off_bound_var]; + const i_t col_end = lp.A.col_start[max_off_bound_var + 1]; + const i_t col_len = col_end - col_start; + const i_t max_potential_rows = 10; + if (col_len > 1) { + std::vector potential_rows; + potential_rows.reserve(col_len); + + const f_t threshold = 1e-4; + for (i_t q = col_start; q < col_end; q++) { + const i_t i = lp.A.i[q]; + const f_t val = lp.A.x[q]; + // Can't use rows that have already been aggregated + if (std::abs(val) > threshold && aggregated_mark[i] == 0) { + potential_rows.push_back(i); + } + if (potential_rows.size() >= max_potential_rows) { break; } + } + work_estimate += 5 * (col_end - col_start); + + if (!potential_rows.empty()) { + std::sort(potential_rows.begin(), potential_rows.end(), [&](i_t a, i_t b) { + return score[a] > score[b]; + }); + work_estimate += 10 * std::log2(10); + + const i_t pivot_row = potential_rows[0]; + + sparse_vector_t pivot_row_inequality(Arow, pivot_row); + f_t pivot_row_rhs = lp.rhs[pivot_row]; + work_estimate += pivot_row_inequality.i.size(); + mir.combine_rows(lp, + Arow, + max_off_bound_var, + pivot_row_inequality, + pivot_row_rhs, + inequality, + inequality_rhs); + aggregated_rows.push_back(pivot_row); + aggregated_mark[pivot_row] = 1; + work_estimate += inequality.i.size() + pivot_row_inequality.i.size(); + } else { + // No potential rows to aggregate + break; + } + } + } + num_aggregated++; // Always increase so the loop terminates + } + } + + if (add_cut) { + // We were successful in generating a cut. + + // Set the score of the aggregated rows to zero + for (i_t row : aggregated_rows) { + score[row] = 0.0; + } + } + + // Clear the aggregated mark + for (i_t row : aggregated_rows) { + aggregated_mark[row] = 0; + } + work_estimate += 2 * aggregated_rows.size(); + // Clear the aggregated rows + aggregated_rows.clear(); + + // Set the score of the current row to zero + score[i] = 0.0; + + // Re-sort the rows by score + // It's possible this could be made more efficient by storing the rows in a data structure + // that allows us to: + // 1. Get the row with the best score + // 2. Get the row with a nonzero in column j that has the best score + // 3. Remove the rows that have been aggregated + // 4. Remove the current row + best_score_last_permutation(score, sorted_indices); + work_estimate += score.size() * std::log2(score.size()); + } +} + +template +void cut_generation_t::generate_gomory_cuts( + const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + csr_matrix_t& Arow, + const std::vector& new_slacks, + const std::vector& var_types, + basis_update_mpf_t& basis_update, + const std::vector& xstar, + const std::vector& basic_list, + const std::vector& nonbasic_list) +{ + tableau_equality_t tableau(lp, basis_update, nonbasic_list); + mixed_integer_rounding_cut_t mir(lp, settings, new_slacks, xstar); + strong_cg_cut_t cg(lp, var_types, xstar); + + for (i_t i = 0; i < lp.num_rows; i++) { + sparse_vector_t inequality(lp.num_cols, 0); + f_t inequality_rhs; + const i_t j = basic_list[i]; + if (var_types[j] != variable_type_t::INTEGER) { continue; } + const f_t x_j = xstar[j]; + if (std::abs(x_j - std::round(x_j)) < settings.integer_tol) { continue; } + i_t tableau_status = tableau.generate_base_equality(lp, + settings, + Arow, + var_types, + basis_update, + xstar, + basic_list, + nonbasic_list, + i, + inequality, + inequality_rhs); + if (tableau_status == 0) { + // Generate a CG cut + const bool generate_cg_cut = settings.strong_chvatal_gomory_cuts != 0; + if (generate_cg_cut) { + // Try to generate a CG cut + sparse_vector_t cg_inequality = inequality; + f_t cg_inequality_rhs = inequality_rhs; + if (fractional_part(inequality_rhs) < 0.5) { + // Multiply by -1 to force the fractional part to be greater than 0.5 + cg_inequality_rhs *= -1; + cg_inequality.negate(); + } + sparse_vector_t cg_cut(lp.num_cols, 0); + f_t cg_cut_rhs; + i_t cg_status = cg.generate_strong_cg_cut( + lp, settings, var_types, cg_inequality, cg_inequality_rhs, xstar, cg_cut, cg_cut_rhs); + if (cg_status == 0) { cut_pool_.add_cut(cut_type_t::CHVATAL_GOMORY, cg_cut, cg_cut_rhs); } + } + + if (settings.mixed_integer_gomory_cuts == 0) { continue; } + + // Given the base inequality, generate a MIR cut + sparse_vector_t cut_A(lp.num_cols, 0); + f_t cut_A_rhs; + i_t mir_status = mir.generate_cut( + inequality, inequality_rhs, lp.upper, lp.lower, var_types, cut_A, cut_A_rhs); + bool A_valid = false; + f_t cut_A_distance = 0.0; + if (mir_status == 0) { + if (cut_A.i.size() == 0) { continue; } + mir.substitute_slacks(lp, Arow, cut_A, cut_A_rhs); + if (cut_A.i.size() == 0) { + A_valid = false; + } else { + // Check that the cut is violated + f_t dot = cut_A.dot(xstar); + f_t cut_norm = cut_A.norm2_squared(); + if (dot >= cut_A_rhs) { continue; } + cut_A_distance = (cut_A_rhs - dot) / std::sqrt(cut_norm); + A_valid = true; + } + } + + // Negate the base inequality + inequality.negate(); + inequality_rhs *= -1; + + sparse_vector_t cut_B(lp.num_cols, 0); + f_t cut_B_rhs; + + mir_status = mir.generate_cut( + inequality, inequality_rhs, lp.upper, lp.lower, var_types, cut_B, cut_B_rhs); + bool B_valid = false; + f_t cut_B_distance = 0.0; + if (mir_status == 0) { + if (cut_B.i.size() == 0) { continue; } + mir.substitute_slacks(lp, Arow, cut_B, cut_B_rhs); + if (cut_B.i.size() == 0) { + B_valid = false; + } else { + // Check that the cut is violated + f_t dot = cut_B.dot(xstar); + f_t cut_norm = cut_B.norm2_squared(); + if (dot >= cut_B_rhs) { continue; } + cut_B_distance = (cut_B_rhs - dot) / std::sqrt(cut_norm); + B_valid = true; + } + } + + if ((cut_A_distance > cut_B_distance) && A_valid) { + cut_pool_.add_cut(cut_type_t::MIXED_INTEGER_GOMORY, cut_A, cut_A_rhs); + } else if (B_valid) { + cut_pool_.add_cut(cut_type_t::MIXED_INTEGER_GOMORY, cut_B, cut_B_rhs); + } + } + } +} + +template +i_t tableau_equality_t::generate_base_equality( + const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + csr_matrix_t& Arow, + const std::vector& var_types, + basis_update_mpf_t& basis_update, + const std::vector& xstar, + const std::vector& basic_list, + const std::vector& nonbasic_list, + i_t i, + sparse_vector_t& inequality, + f_t& inequality_rhs) +{ + // Let's look for Gomory cuts + const i_t j = basic_list[i]; + if (var_types[j] != variable_type_t::INTEGER) { return -1; } + const f_t x_j = xstar[j]; + if (std::abs(x_j - std::round(x_j)) < settings.integer_tol) { return -1; } +#ifdef PRINT_CUT_INFO + settings_.log.printf("Generating cut for variable %d relaxed value %e row %d\n", j, x_j, i); +#endif + + // Solve B^T u_bar = e_i + sparse_vector_t e_i(lp.num_rows, 1); + e_i.i[0] = i; + e_i.x[0] = 1.0; + sparse_vector_t u_bar(lp.num_rows, 0); + basis_update.b_transpose_solve(e_i, u_bar); + +#ifdef CHECK_B_TRANSPOSE_SOLVE + std::vector u_bar_dense(lp.num_rows); + u_bar.to_dense(u_bar_dense); + + std::vector BTu_bar(lp.num_rows); + b_transpose_multiply(lp, basic_list, u_bar_dense, BTu_bar); + for (i_t k = 0; k < lp.num_rows; k++) { + if (k == i) { + settings.log.printf("BTu_bar %d error %e\n", k, std::abs(BTu_bar[k] - 1.0)); + if (std::abs(BTu_bar[k] - 1.0) > 1e-10) { + settings.log.printf("BTu_bar[%d] = %e i %d\n", k, BTu_bar[k], i); + assert(false); + } + } else { + settings.log.printf("BTu_bar %d error %e\n", k, std::abs(BTu_bar[k])); + if (std::abs(BTu_bar[k]) > 1e-10) { + settings.log.printf("BTu_bar[%d] = %e i %d\n", k, BTu_bar[k], i); + assert(false); + } + } + } +#endif + + // Compute a_bar = N^T u_bar + // TODO: This is similar to a function in phase2 of dual simplex. See if it can be reused. + const i_t nz_ubar = u_bar.i.size(); + std::vector abar_indices; + abar_indices.reserve(nz_ubar); + for (i_t k = 0; k < nz_ubar; k++) { + const i_t ii = u_bar.i[k]; + const f_t u_bar_i = u_bar.x[k]; + const i_t row_start = Arow.row_start[ii]; + const i_t row_end = Arow.row_start[ii + 1]; + for (i_t p = row_start; p < row_end; p++) { + const i_t jj = Arow.j[p]; + if (nonbasic_mark_[jj] == 1) { + const f_t val = u_bar_i * Arow.x[p]; + const f_t y = val - c_workspace_[jj]; + const f_t t = x_workspace_[jj] + y; + c_workspace_[jj] = (t - x_workspace_[jj]) - y; + x_workspace_[jj] = t; + if (!x_mark_[jj]) { + x_mark_[jj] = 1; + abar_indices.push_back(jj); + } + } + } + } + // TODO: abar has lots of small coefficients. Double check that + // we do not accidently create a base (in)equality + // that cuts off an integer solution, when we drop the small coefficients. + + i_t small_coeff = 0; + const f_t drop_tol = 1e-12; + const bool drop_coefficients = true; + sparse_vector_t a_bar(lp.num_cols, 0); + a_bar.i.reserve(abar_indices.size() + 1); + a_bar.x.reserve(abar_indices.size() + 1); + for (i_t k = 0; k < abar_indices.size(); k++) { + const i_t jj = abar_indices[k]; + if (drop_coefficients && std::abs(x_workspace_[jj]) < drop_tol) { + small_coeff++; + } else { + a_bar.i.push_back(jj); + a_bar.x.push_back(x_workspace_[jj]); + } + } + const bool verbose = false; + if (verbose && small_coeff > 0) { settings.log.printf("Small coeff dropped %d\n", small_coeff); } + + // Clear the workspace + for (i_t jj : abar_indices) { + x_workspace_[jj] = 0.0; + x_mark_[jj] = 0; + c_workspace_[jj] = 0.0; + } + abar_indices.clear(); + + // We should now have the base inequality + // x_j + a_bar^T x_N >= b_bar_i + // We add x_j into a_bar so that everything is in a single sparse_vector_t + a_bar.i.push_back(j); + a_bar.x.push_back(1.0); + + // Check that the tableau equality is satisfied + const f_t tableau_tol = 1e-6; + f_t a_bar_dot_xstar = a_bar.dot(xstar); + if (std::abs(a_bar_dot_xstar - b_bar_[i]) > tableau_tol) { + settings.log.debug("bad tableau equality. error %e\n", std::abs(a_bar_dot_xstar - b_bar_[i])); + return -1; + } + + // We have that x_j + a_bar^T x_N == b_bar_i + // So x_j + a_bar^T x_N >= b_bar_i + // And x_j + a_bar^T x_N <= b_bar_i + // Or -x_j - a_bar^T x_N >= -b_bar_i + + // Skip cuts that are shallow + const f_t shallow_tol = 1e-2; + if (std::abs(x_j - std::round(x_j)) < shallow_tol) { + // Skip cuts where integer variable has small fractional part + return -1; + } + + const f_t f_val = b_bar_[i] - std::floor(b_bar_[i]); + if (f_val < 0.01 || f_val > 0.99) { + // Skip cuts with rhs has small fractional part + return -1; + } + +#ifdef PRINT_BASE_INEQUALITY + // Print out the base inequality + for (i_t k = 0; k < a_bar.i.size(); k++) { + const i_t jj = a_bar.i[k]; + const f_t aj = a_bar.x[k]; + settings_.log.printf("a_bar[%d] = %e\n", k, aj); + } + settings_.log.printf("b_bar[%d] = %e\n", i, b_bar[i]); +#endif + + inequality = a_bar; + inequality_rhs = b_bar_[i]; + + return 0; +} + +template +mixed_integer_rounding_cut_t::mixed_integer_rounding_cut_t( + const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + const std::vector& new_slacks, + const std::vector& xstar) + : num_vars_(lp.num_cols), + settings_(settings), + x_workspace_(num_vars_, 0.0), + x_mark_(num_vars_, 0), + has_lower_(num_vars_, 0), + has_upper_(num_vars_, 0), + is_slack_(num_vars_, 0), + slack_rows_(num_vars_, 0), + bound_info_(num_vars_, 0) +{ + for (i_t j : new_slacks) { + is_slack_[j] = 1; + const i_t col_start = lp.A.col_start[j]; + const i_t i = lp.A.i[col_start]; + slack_rows_[j] = i; + assert(std::abs(lp.A.x[col_start]) == 1.0); + } + + needs_complement_ = false; + for (i_t j = 0; j < num_vars_; j++) { + if (lp.lower[j] < 0) { + settings_.log.debug("Variable %d has negative lower bound %e\n", j, lp.lower[j]); + } + const f_t uj = lp.upper[j]; + const f_t lj = lp.lower[j]; + const f_t xstar_j = xstar[j]; + if (uj < inf) { + if (uj - xstar_j <= xstar_j - lj) { + has_upper_[j] = 1; + bound_info_[j] = 1; + needs_complement_ = true; + } else if (lj != 0.0) { + has_lower_[j] = 1; + bound_info_[j] = -1; + needs_complement_ = true; + } + continue; + } + + if (lj > -inf && lj != 0.0) { + has_lower_[j] = 1; + bound_info_[j] = -1; + needs_complement_ = true; + } + } +} + +template +void mixed_integer_rounding_cut_t::to_nonnegative(const lp_problem_t& lp, + sparse_vector_t& inequality, + f_t& rhs) +{ + const i_t nz = inequality.i.size(); + for (i_t k = 0; k < nz; k++) { + const i_t j = inequality.i[k]; + const f_t aj = inequality.x[k]; + if (bound_info_[j] == -1) { + // v_j = x_j - l_j, v_j >= 0 + // x_j = v_j + l_j + // sum_{k != j} a_k x_j + a_j x_j <= beta + // sum_{k != j} a_k x_j + a_j (v_j + l_j) <= beta + // sum_{k != j} a_k x_j + a_j v_j <= beta - a_j l_j + const f_t lj = lp.lower[j]; + rhs -= aj * lj; + } else if (bound_info_[j] == 1) { + // w_j = u_j - x_j, w_j >= 0 + // x_j = u_j - w_j + // sum_{k != j} a_k x_k + a_j x_j <= beta + // sum_{k != j} a_k x_k + a_j (u_j - w_j) <= beta + // sum_{k != j} a_k x_k - a_j w_j <= beta - a_j u_j + const f_t uj = lp.upper[j]; + inequality.x[k] *= -1.0; + rhs -= aj * uj; + } + } +} + +template +void mixed_integer_rounding_cut_t::relaxation_to_nonnegative( + const lp_problem_t& lp, + const std::vector& xstar, + std::vector& xstar_nonnegative) +{ + xstar_nonnegative = xstar; + const i_t n = lp.num_cols; + for (i_t j = 0; j < n; ++j) { + if (bound_info_[j] == -1) { + // v_j = x_j - l_j + const f_t lj = lp.lower[j]; + xstar_nonnegative[j] -= lj; + } else if (bound_info_[j] == 1) { + // w_j = u_j - x_j + const f_t uj = lp.upper[j]; + xstar_nonnegative[j] = uj - xstar_nonnegative[j]; + } + } +} + +template +void mixed_integer_rounding_cut_t::to_original(const lp_problem_t& lp, + sparse_vector_t& inequality, + f_t& rhs) +{ + const i_t nz = inequality.i.size(); + for (i_t k = 0; k < nz; k++) { + const i_t j = inequality.i[k]; + const f_t dj = inequality.x[k]; + if (bound_info_[j] == -1) { + // v_j = x_j - l_j, v_j >= 0 + // sum_{k != j} d_k x_k + d_j v_j >= beta + // sum_{k != j} d_k x_k + d_j (x_j - l_j) >= beta + // sum_{k != j} d_k x_k + d_j x_j >= beta + d_j l_j + const f_t lj = lp.lower[j]; + rhs += dj * lj; + } else if (bound_info_[j] == 1) { + // w_j = u_j - x_j, w_j >= 0 + // sum_{k != j} d_k x_k + d_j w_j >= beta + // sum_{k != j} d_k x_k + d_j (u_j - x_j) >= beta + // sum_{k != j} d_k x_k - d_j x_j >= beta - d_j u_j + const f_t uj = lp.upper[j]; + inequality.x[k] *= -1.0; + rhs -= dj * uj; + } + } +} + +template +void mixed_integer_rounding_cut_t::remove_small_coefficients( + const std::vector& lower_bounds, + const std::vector& upper_bounds, + sparse_vector_t& cut, + f_t& cut_rhs) +{ + const i_t nz = cut.i.size(); + i_t removed = 0; + for (i_t k = 0; k < cut.i.size(); k++) { + const i_t j = cut.i[k]; + + // Check for small coefficients + const f_t aj = cut.x[k]; + if (std::abs(aj) < 1e-6) { + if (aj >= 0.0 && upper_bounds[j] < inf) { + // Move this to the right-hand side + cut_rhs -= aj * upper_bounds[j]; + cut.x[k] = 0.0; + removed++; + } else if (aj <= 0.0 && lower_bounds[j] > -inf) { + cut_rhs += aj * lower_bounds[j]; + cut.x[k] = 0.0; + removed++; + continue; + } else { + } + } + } + + if (removed > 0) { + sparse_vector_t new_cut(cut.n, 0); + cut.squeeze(new_cut); + cut = new_cut; + } +} + +template +i_t mixed_integer_rounding_cut_t::generate_cut_nonnegative( + const sparse_vector_t& a, + f_t beta, + const std::vector& var_types, + sparse_vector_t& cut, + f_t& cut_rhs) +{ + auto f = [](f_t q_1, f_t q_2) -> f_t { + f_t q_1_hat = q_1 - std::floor(q_1); + f_t q_2_hat = q_2 - std::floor(q_2); + return std::min(q_1_hat, q_2_hat) + q_2_hat * std::floor(q_1); + }; + + auto h = [](f_t q) -> f_t { return std::max(q, 0.0); }; + + std::vector cut_indices; + cut_indices.reserve(a.i.size()); + f_t R = (beta - std::floor(beta)) * std::ceil(beta); + + for (i_t k = 0; k < a.i.size(); k++) { + const i_t jj = a.i[k]; + f_t aj = a.x[k]; + if (var_types[jj] == variable_type_t::INTEGER) { + x_workspace_[jj] += f(aj, beta); + if (!x_mark_[jj] && x_workspace_[jj] != 0.0) { + x_mark_[jj] = 1; + cut_indices.push_back(jj); + } + } else { + x_workspace_[jj] += h(aj); + if (!x_mark_[jj] && x_workspace_[jj] != 0.0) { + x_mark_[jj] = 1; + cut_indices.push_back(jj); + } + } + } + + cut.i.reserve(cut_indices.size()); + cut.x.reserve(cut_indices.size()); + cut.i.clear(); + cut.x.clear(); + for (i_t k = 0; k < cut_indices.size(); k++) { + const i_t j = cut_indices[k]; + cut.i.push_back(j); + cut.x.push_back(x_workspace_[j]); + } + + // Clear the workspace + for (i_t jj : cut_indices) { + x_workspace_[jj] = 0.0; + x_mark_[jj] = 0; + } + +#ifdef CHECK_WORKSPACE + for (i_t j = 0; j < x_workspace_.size(); j++) { + if (x_workspace_[j] != 0.0) { + printf("After generate_cut: Dirty x_workspace_[%d] = %e\n", j, x_workspace_[j]); + assert(x_workspace_[j] == 0.0); + } + if (x_mark_[j] != 0) { + printf("After generate_cut: Dirty x_mark_[%d] = %d\n", j, x_mark_[j]); + assert(x_mark_[j] == 0); + } + } +#endif + + // The new cut is: g'*x >= R + // But we want to have it in the form h'*x <= b + cut.sort(); + + cut_rhs = R; + +#ifdef CHECK_REPEATED_INDICES + // Check for repeated indicies + std::vector check(num_vars_, 0); + for (i_t p = 0; p < cut.i.size(); p++) { + if (check[cut.i[p]] != 0) { + printf("repeated index in generated cut\n"); + assert(check[cut.i[p]] == 0); + } + check[cut.i[p]] = 1; + } +#endif + + if (cut.i.size() == 0) { return -1; } + + return 0; +} + +template +i_t mixed_integer_rounding_cut_t::generate_cut( + const sparse_vector_t& a, + f_t beta, + const std::vector& upper_bounds, + const std::vector& lower_bounds, + const std::vector& var_types, + sparse_vector_t& cut, + f_t& cut_rhs) +{ +#ifdef CHECK_WORKSPACE + for (i_t j = 0; j < x_workspace_.size(); j++) { + if (x_workspace_[j] != 0.0) { + printf("Before generate_cut: Dirty x_workspace_[%d] = %e\n", j, x_workspace_[j]); + printf("num_vars_ %d\n", num_vars_); + printf("x_workspace_.size() %ld\n", x_workspace_.size()); + assert(x_workspace_[j] == 0.0); + } + if (x_mark_[j] != 0) { + printf("Before generate_cut: Dirty x_mark_[%d] = %d\n", j, x_mark_[j]); + assert(x_mark_[j] == 0); + } + } +#endif + + auto f = [](f_t q_1, f_t q_2) -> f_t { + f_t q_1_hat = q_1 - std::floor(q_1); + f_t q_2_hat = q_2 - std::floor(q_2); + return std::min(q_1_hat, q_2_hat) + q_2_hat * std::floor(q_1); + }; + + auto h = [](f_t q) -> f_t { return std::max(q, 0.0); }; + + std::vector cut_indices; + cut_indices.reserve(a.i.size()); + f_t R; + if (!needs_complement_) { + R = (beta - std::floor(beta)) * std::ceil(beta); + + for (i_t k = 0; k < a.i.size(); k++) { + const i_t jj = a.i[k]; + f_t aj = a.x[k]; + if (var_types[jj] == variable_type_t::INTEGER) { + x_workspace_[jj] += f(aj, beta); + if (!x_mark_[jj] && x_workspace_[jj] != 0.0) { + x_mark_[jj] = 1; + cut_indices.push_back(jj); + } + } else { + x_workspace_[jj] += h(aj); + if (!x_mark_[jj] && x_workspace_[jj] != 0.0) { + x_mark_[jj] = 1; + cut_indices.push_back(jj); + } + } + } + } else { + // Compute r + f_t r = beta; + for (i_t k = 0; k < a.i.size(); k++) { + const i_t jj = a.i[k]; + if (has_upper_[jj]) { + const f_t uj = upper_bounds[jj]; + r -= uj * a.x[k]; + continue; + } + if (has_lower_[jj]) { + const f_t lj = lower_bounds[jj]; + r -= lj * a.x[k]; + } + } + + // Compute R + R = std::ceil(r) * (r - std::floor(r)); + for (i_t k = 0; k < a.i.size(); k++) { + const i_t jj = a.i[k]; + const f_t aj = a.x[k]; + if (has_upper_[jj]) { + const f_t uj = upper_bounds[jj]; + if (var_types[jj] == variable_type_t::INTEGER) { + R -= f(-aj, r) * uj; + } else { + R -= h(-aj) * uj; + } + } else if (has_lower_[jj]) { + const f_t lj = lower_bounds[jj]; + if (var_types[jj] == variable_type_t::INTEGER) { + R += f(aj, r) * lj; + } else { + R += h(aj) * lj; + } + } + } + + // Compute the cut coefficients + for (i_t k = 0; k < a.i.size(); k++) { + const i_t jj = a.i[k]; + const f_t aj = a.x[k]; + if (has_upper_[jj]) { + if (var_types[jj] == variable_type_t::INTEGER) { + // Upper intersect I + x_workspace_[jj] -= f(-aj, r); + if (!x_mark_[jj] && x_workspace_[jj] != 0.0) { + x_mark_[jj] = 1; + cut_indices.push_back(jj); + } + } else { + // Upper intersect C + f_t h_j = h(-aj); + if (h_j != 0.0) { + x_workspace_[jj] -= h_j; + if (!x_mark_[jj]) { + x_mark_[jj] = 1; + cut_indices.push_back(jj); + } + } + } + } else if (var_types[jj] == variable_type_t::INTEGER) { + // I \ Upper + x_workspace_[jj] += f(aj, r); + if (!x_mark_[jj] && x_workspace_[jj] != 0.0) { + x_mark_[jj] = 1; + cut_indices.push_back(jj); + } + } else { + // C \ Upper + f_t h_j = h(aj); + if (h_j != 0.0) { + x_workspace_[jj] += h_j; + if (!x_mark_[jj]) { + x_mark_[jj] = 1; + cut_indices.push_back(jj); + } + } + } + } + } + + cut.i.reserve(cut_indices.size()); + cut.x.reserve(cut_indices.size()); + cut.i.clear(); + cut.x.clear(); + for (i_t k = 0; k < cut_indices.size(); k++) { + const i_t jj = cut_indices[k]; + + // Check for small coefficients + const f_t aj = x_workspace_[jj]; + if (std::abs(aj) < 1e-6) { + if (aj >= 0.0 && upper_bounds[jj] < inf) { + // Move this to the right-hand side + R -= aj * upper_bounds[jj]; + continue; + } else if (aj <= 0.0 && lower_bounds[jj] > -inf) { + R += aj * lower_bounds[jj]; + continue; + } else { + } + } + cut.i.push_back(jj); + cut.x.push_back(x_workspace_[jj]); + } + + // Clear the workspace + for (i_t jj : cut_indices) { + x_workspace_[jj] = 0.0; + x_mark_[jj] = 0; + } + +#ifdef CHECK_WORKSPACE + for (i_t j = 0; j < x_workspace_.size(); j++) { + if (x_workspace_[j] != 0.0) { + printf("After generate_cut: Dirty x_workspace_[%d] = %e\n", j, x_workspace_[j]); + assert(x_workspace_[j] == 0.0); + } + if (x_mark_[j] != 0) { + printf("After generate_cut: Dirty x_mark_[%d] = %d\n", j, x_mark_[j]); + assert(x_mark_[j] == 0); + } + } +#endif + + // The new cut is: g'*x >= R + // But we want to have it in the form h'*x <= b + cut.sort(); + + cut_rhs = R; + +#ifdef CHECK_REPEATED_INDICES + // Check for repeated indicies + std::vector check(num_vars_, 0); + for (i_t p = 0; p < cut.i.size(); p++) { + if (check[cut.i[p]] != 0) { + printf("repeated index in generated cut\n"); + assert(check[cut.i[p]] == 0); + } + check[cut.i[p]] = 1; + } +#endif + + if (cut.i.size() == 0) { return -1; } + + return 0; +} + +template +void mixed_integer_rounding_cut_t::substitute_slacks(const lp_problem_t& lp, + csr_matrix_t& Arow, + sparse_vector_t& cut, + f_t& cut_rhs) +{ + // Remove slacks from the cut + // So that the cut is only over the original variables + bool found_slack = false; + i_t cut_nz = 0; + std::vector cut_indices; + cut_indices.reserve(cut.i.size()); + +#ifdef CHECK_WORKSPACE + for (i_t j = 0; j < x_workspace_.size(); j++) { + if (x_workspace_[j] != 0.0) { + printf("Begin Dirty x_workspace_[%d] = %e\n", j, x_workspace_[j]); + assert(x_workspace_[j] == 0.0); + } + if (x_mark_[j] != 0) { + printf("Begin Dirty x_mark_[%d] = %d\n", j, x_mark_[j]); + assert(x_mark_[j] == 0); + } + } +#endif + + for (i_t k = 0; k < cut.i.size(); k++) { + const i_t j = cut.i[k]; + const f_t cj = cut.x[k]; + if (is_slack_[j]) { + found_slack = true; + const i_t slack_start = lp.A.col_start[j]; +#ifdef CHECK_SLACKS + const i_t slack_end = lp.A.col_start[j + 1]; + const i_t slack_len = slack_end - slack_start; + if (slack_len != 1) { + printf("Slack %d has %d nzs in colum\n", j, slack_len); + assert(slack_len == 1); + } +#endif + const f_t alpha = lp.A.x[slack_start]; +#ifdef CHECK_SLACKS + if (std::abs(alpha) != 1.0) { + printf("Slack %d has non-unit coefficient %e\n", j, alpha); + assert(std::abs(alpha) == 1.0); + } +#endif + + // Do the substitution + // Slack variable s_j participates in row i of the constraint matrix + // Row i is of the form: + // sum_{k != j} A(i, k) * x_k + alpha * s_j = rhs_i + // where alpha = +1/-1 + /// So we have that + // s_j = (rhs_i - sum_{k != j} A(i, k) * x_k)/alpha + + // Our cut is of the form: + // sum_{k != j} C(k) * x_k + C(j) * s_j >= cut_rhs + // So the cut becomes + // sum_{k != j} C(k) * x_k + C(j)/alpha * (rhs_i - sum_{h != j} A(i, h) * x_h) >= cut_rhs + // This is equivalent to: + // sum_{k != j} C(k) * x_k + sum_{h != j} -C(j)/alpha * A(i, h) * x_h >= cut_rhs - C(j)/alpha + // * rhs_i + const i_t i = slack_rows_[j]; + cut_rhs -= cj * lp.rhs[i] / alpha; + const i_t row_start = Arow.row_start[i]; + const i_t row_end = Arow.row_start[i + 1]; + for (i_t q = row_start; q < row_end; q++) { + const i_t h = Arow.j[q]; + if (h != j) { + const f_t aih = Arow.x[q]; + x_workspace_[h] -= cj * aih / alpha; + if (!x_mark_[h]) { + x_mark_[h] = 1; + cut_indices.push_back(h); + cut_nz++; + } + } else { + const f_t aij = Arow.x[q]; + if (std::abs(aij) != 1.0) { + settings_.log.printf( + "Slack row %d has non-unit coefficient %e for variable %d\n", i, aij, j); + assert(std::abs(aij) == 1.0); + } + } + } + + } else { + x_workspace_[j] += cj; + if (!x_mark_[j]) { + x_mark_[j] = 1; + cut_indices.push_back(j); + cut_nz++; + } + } + } + + if (found_slack) { + cut.i.reserve(cut_nz); + cut.x.reserve(cut_nz); + cut.i.clear(); + cut.x.clear(); + + for (i_t k = 0; k < cut_nz; k++) { + const i_t j = cut_indices[k]; + + // Check for small coefficients + const f_t aj = x_workspace_[j]; + if (std::abs(aj) < 1e-6) { + if (aj >= 0.0 && lp.upper[j] < inf) { + // Move this to the right-hand side + cut_rhs -= aj * lp.upper[j]; + continue; + } else if (aj <= 0.0 && lp.lower[j] > -inf) { + cut_rhs += aj * lp.lower[j]; + continue; + } else { + } + } + + cut.i.push_back(j); + cut.x.push_back(x_workspace_[j]); + } + // Sort the cut + cut.sort(); + } + + // Clear the workspace + for (i_t jj : cut_indices) { + x_workspace_[jj] = 0.0; + x_mark_[jj] = 0; + } + +#ifdef CHECK_WORKSPACE + for (i_t j = 0; j < x_workspace_.size(); j++) { + if (x_workspace_[j] != 0.0) { + printf("End Dirty x_workspace_[%d] = %e\n", j, x_workspace_[j]); + assert(x_workspace_[j] == 0.0); + } + if (x_mark_[j] != 0) { + printf("End Dirty x_mark_[%d] = %d\n", j, x_mark_[j]); + assert(x_mark_[j] == 0); + } + } +#endif +} + +template +f_t mixed_integer_rounding_cut_t::compute_violation(const sparse_vector_t& cut, + f_t cut_rhs, + const std::vector& xstar) +{ + f_t dot = cut.dot(xstar); + f_t cut_violation = cut_rhs - dot; + return cut_violation; +} + +template +void mixed_integer_rounding_cut_t::combine_rows( + const lp_problem_t& lp, + csr_matrix_t& Arow, + i_t xj, + const sparse_vector_t& pivot_row, + f_t pivot_row_rhs, + sparse_vector_t& inequality, + f_t& inequality_rhs) +{ +#ifdef CHECK_WORKSPACE + for (i_t k = 0; k < x_workspace_.size(); k++) { + if (x_workspace_[k] != 0.0) { + printf("Dirty x_workspace_[%d] = %e\n", k, x_workspace_[k]); + assert(x_workspace_[k] == 0.0); + } + if (x_mark_[k] != 0) { + printf("Dirty x_mark_[%d] = %d\n", k, x_mark_[k]); + assert(x_mark_[k] == 0); + } + } +#endif + + indices_.clear(); + indices_.reserve(pivot_row.i.size() + inequality.i.size()); + + // Find the coefficient associated with variable xj in the pivot row + f_t a_l_j = 0.0; + for (i_t k = 0; k < pivot_row.i.size(); k++) { + const i_t j = pivot_row.i[k]; + if (j == xj) { + a_l_j = pivot_row.x[k]; + break; + } + } + + if (a_l_j == 0) { return; } + + f_t a_i_j = 0.0; + + i_t nz = 0; + // Store the inequality in the workspace + // and save the coefficient associated with variable xj + for (i_t k = 0; k < inequality.i.size(); k++) { + const i_t j = inequality.i[k]; + if (j != xj) { + x_workspace_[j] = inequality.x[k]; + x_mark_[j] = 1; + indices_.push_back(j); + nz++; + } else { + a_i_j = inequality.x[k]; + } + } + + f_t pivot_value = a_i_j / a_l_j; + // Adjust the rhs of the inequality + inequality_rhs -= pivot_value * pivot_row_rhs; + + // Adjust the coefficients of the inequality + // based on the nonzeros in the pivot row + for (i_t k = 0; k < pivot_row.i.size(); k++) { + const i_t j = pivot_row.i[k]; + if (j != xj) { + x_workspace_[j] -= pivot_value * pivot_row.x[k]; + if (!x_mark_[j]) { + x_mark_[j] = 1; + indices_.push_back(j); + nz++; + } + } + } + + // Store the new inequality + inequality.i.resize(nz); + inequality.x.resize(nz); + for (i_t k = 0; k < nz; k++) { + inequality.i[k] = indices_[k]; + inequality.x[k] = x_workspace_[indices_[k]]; + } + +#ifdef CHECK_REPEATED_INDICES + // Check for repeated indices + std::vector check(num_vars_, 0); + for (i_t k = 0; k < inequality.i.size(); k++) { + if (check[inequality.i[k]] == 1) { + printf("repeated index\n"); + assert(check[inequality.i[k]] == 0); + } + check[inequality.i[k]] = 1; + } +#endif + + // Clear the workspace + for (i_t j : indices_) { + x_workspace_[j] = 0.0; + x_mark_[j] = 0; + } + indices_.clear(); +} + +template +strong_cg_cut_t::strong_cg_cut_t(const lp_problem_t& lp, + const std::vector& var_types, + const std::vector& xstar) + : transformed_variables_(lp.num_cols, 0) +{ + // Determine the substition for the integer variables + for (i_t j = 0; j < lp.num_cols; j++) { + if (var_types[j] == variable_type_t::INTEGER) { + const f_t l_j = lp.lower[j]; + const f_t u_j = lp.upper[j]; + if (l_j != 0.0) { + // We need to transform the variable + // Check the distance to each bound + const f_t dist_to_lower = std::max(0.0, xstar[j] - l_j); + const f_t dist_to_upper = std::max(0.0, u_j - xstar[j]); + if (dist_to_upper >= dist_to_lower || u_j >= inf) { + // We are closer to the lower bound. + transformed_variables_[j] = -1; + } else if (u_j < inf) { + // We are closer to the finite upper bound + transformed_variables_[j] = 1; + } + } + } + } +} + +template +i_t strong_cg_cut_t::remove_continuous_variables_integers_nonnegative( + const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + const std::vector& var_types, + sparse_vector_t& inequality, + f_t& inequality_rhs) +{ + const bool verbose = false; + // Count the number of continuous variables in the inequality + i_t num_continuous = 0; + const i_t nz = inequality.i.size(); + for (i_t k = 0; k < nz; k++) { + const i_t j = inequality.i[k]; + if (var_types[j] == variable_type_t::CONTINUOUS) { num_continuous++; } + } + + if (verbose) { settings.log.printf("num_continuous %d\n", num_continuous); } + // We assume the inequality is of the form sum_j a_j x_j <= rhs + + for (i_t k = 0; k < nz; k++) { + const i_t j = inequality.i[k]; + const f_t l_j = lp.lower[j]; + const f_t u_j = lp.upper[j]; + const f_t a_j = inequality.x[k]; + if (var_types[j] == variable_type_t::CONTINUOUS) { + if (a_j == 0.0) { continue; } + + if (a_j > 0.0 && l_j > -inf) { + // v_j = x_j - l_j >= 0 + // x_j = v_j + l_j + // sum_{k != j} a_k x_k + a_j x_j <= rhs + // sum_{k != j} a_k x_k + a_j (v_j + l_j) <= rhs + // sum_{k != j} a_k x_k + a_j v_j <= rhs - a_j l_j + inequality_rhs -= a_j * l_j; + transformed_variables_[j] = -1; + + // We now have a_j * v_j with a_j, v_j >= 0 + // So we have sum_{k != j} a_k x_k <= sum_{k != j} a_k x_k + a_j v_j <= rhs - a_j l_j + // So we can now drop the continuous variable v_j + inequality.x[k] = 0.0; + + } else if (a_j < 0.0 && u_j < inf) { + // w_j = u_j - x_j >= 0 + // x_j = u_j - w_j + // sum_{k != j} a_k x_k + a_j x_j <= rhs + // sum_{k != j} a_k x_k + a_j (u_j - w_j) <= rhs + // sum_{k != j} a_k x_k - a_j w_j <= rhs - a_j u_j + inequality_rhs -= a_j * u_j; + transformed_variables_[j] = 1; + + // We now have a_j * w_j with a_j, w_j >= 0 + // So we have sum_{k != j} a_k x_k <= sum_{k != j} a_k x_k + a_j w_j <= rhs - a_j u_j + // So we can now drop the continuous variable w_j + inequality.x[k] = 0.0; + } else { + // We can't keep the coefficient of the continuous variable positive + // This means we can't eliminate the continuous variable + if (verbose) { settings.log.printf("x%d ak: %e lo: %e up: %e\n", j, a_j, l_j, u_j); } + return -1; + } + } else { + // The variable is integer. We just need to ensure it is nonnegative + if (transformed_variables_[j] == -1) { + // We are closer to the lower bound. + // v_j = x_j - l_j >= 0 + // x_j = v_j + l_j + // sum_{k != j} a_k x_k + a_j x_j <= rhs + // sum_{k != j} a_k x_k + a_j (v_j + l_j) <= rhs + // sum_{k != j} a_k x_k + a_j v_j <= rhs - a_j l_j + inequality_rhs -= a_j * l_j; + } else if (transformed_variables_[j] == 1) { + // We are closer to the finite upper bound + // w_j = u_j - x_j >= 0 + // x_j = u_j - w_j + // sum_{k != j} a_k x_k + a_j x_j <= rhs + // sum_{k != j} a_k x_k + a_j (u_j - w_j) <= rhs + // sum_{k != j} a_k x_k - a_j w_j <= rhs - a_j u_j + inequality_rhs -= a_j * u_j; + inequality.x[k] *= -1.0; + } + } + } + + // Squeeze out the zero coefficents + sparse_vector_t new_inequality(inequality.n, 0); + inequality.squeeze(new_inequality); + inequality = new_inequality; + return 0; +} + +template +void strong_cg_cut_t::to_original_integer_variables(const lp_problem_t& lp, + sparse_vector_t& cut, + f_t& cut_rhs) +{ + // We expect a cut of the form sum_j a_j y_j <= rhs + // where y_j >= 0 is a transformed variable + // We need to convert it back into a cut on the original variables + + for (i_t k = 0; k < cut.i.size(); k++) { + const i_t j = cut.i[k]; + const f_t a_j = cut.x[k]; + if (transformed_variables_[j] == -1) { + // sum_{k != j} a_k x_k + a_j v_j <= rhs + // v_j = x_j - l_j >= 0, + // sum_{k != j} a_k x_k + a_j (x_j - l_j) <= rhs + // sum_{k != j} a_k x_k + a_j x_j <= rhs + a_j l_j + cut_rhs += a_j * lp.lower[j]; + } else if (transformed_variables_[j] == 1) { + // sum_{k != j} a_k x_k + a_j w_j <= rhs + // w_j = u_j - x_j >= 0 + // sum_{k != j} a_k x_k + a_j (u_j - x_j) <= rhs + // sum_{k != j} a_k x_k - a_j x_j <= rhs - a_j u_j + cut_rhs -= a_j * lp.upper[j]; + cut.x[k] *= -1.0; + } + } +} + +template +i_t strong_cg_cut_t::generate_strong_cg_cut_integer_only( + const simplex_solver_settings_t& settings, + const std::vector& var_types, + const sparse_vector_t& inequality, + f_t inequality_rhs, + sparse_vector_t& cut, + f_t& cut_rhs) +{ + // We expect an inequality of the form sum_j a_j x_j <= rhs + // where all the variables x_j are integer and nonnegative + + // We then apply the CG cut: + // sum_j floor(a_j) x_j <= floor(rhs) + cut.i.reserve(inequality.i.size()); + cut.x.reserve(inequality.i.size()); + cut.i.clear(); + cut.x.clear(); + + f_t a_0 = inequality_rhs; + f_t f_a_0 = fractional_part(a_0); + + if (f_a_0 == 0.0) { + // f(a_0) == 0.0 so we do a weak CG cut + cut.i.reserve(inequality.i.size()); + cut.x.reserve(inequality.i.size()); + cut.i.clear(); + cut.x.clear(); + for (i_t k = 0; k < inequality.i.size(); k++) { + const i_t j = inequality.i[k]; + const f_t a_j = inequality.x[k]; + if (var_types[j] == variable_type_t::INTEGER) { + cut.i.push_back(j); + cut.x.push_back(std::floor(a_j)); + } else { + return -1; + } + } + cut_rhs = std::floor(inequality_rhs); + } else { + return generate_strong_cg_cut_helper( + inequality.i, inequality.x, inequality_rhs, var_types, cut, cut_rhs); + } + return 0; +} + +template +i_t strong_cg_cut_t::generate_strong_cg_cut_helper( + const std::vector& indicies, + const std::vector& coefficients, + f_t rhs, + const std::vector& var_types, + sparse_vector_t& cut, + f_t& cut_rhs) +{ + const bool verbose = false; + const i_t nz = indicies.size(); + const f_t f_a_0 = fractional_part(rhs); + + const f_t min_fractional_part = 1e-2; + if (f_a_0 < min_fractional_part) { return -1; } + if (f_a_0 > 1 - min_fractional_part) { return -1; } + + // We will try to generat a strong CG cut. + // Find the unique integer k such that + // 1/(k+1) <= f(a_0) < 1/k + const f_t k_upper = 1.0 / f_a_0; + i_t k = static_cast(std::ceil(k_upper)) - 1; + + const f_t alpha = 1.0 - f_a_0; + f_t lower = 1.0 / static_cast(k + 1); + f_t upper = 1.0 / static_cast(k); + if (verbose) { printf("f_a_0 %e lower %e upper %e alpha %e\n", f_a_0, lower, upper, alpha); } + if (f_a_0 >= lower && f_a_0 < upper) { + cut.i.reserve(nz); + cut.x.reserve(nz); + cut.i.clear(); + cut.x.clear(); + for (i_t q = 0; q < nz; q++) { + const i_t j = indicies[q]; + const f_t a_j = coefficients[q]; + if (var_types[j] == variable_type_t::INTEGER) { + const f_t f_a_j = fractional_part(a_j); + const f_t tol = 1e-4; + if (f_a_j <= f_a_0 + tol) { + cut.i.push_back(j); + cut.x.push_back((k + 1.0) * std::floor(a_j)); + if (verbose) { printf("j %d a_j %e f_a_j %e k %d\n", j, a_j, f_a_j, k); } + } else { + // Find p such that p <= k * f(a_j) < p + 1 + i_t p = static_cast(std::floor(k * f_a_j)); + // If f(a_j) > f(a_0) + p /k (1 - f(a_0)) then we can increase the cofficient by 1 + const f_t rhs_j = f_a_0 + static_cast(p) / static_cast(k) * alpha; + const i_t coeff = (k + 1) * static_cast(std::floor(a_j)) + p; + if (f_a_j > rhs_j + tol) { + cut.i.push_back(j); + cut.x.push_back(static_cast(coeff + 1)); + } else { + cut.i.push_back(j); + cut.x.push_back(static_cast(coeff)); + } + } + } else { + return -1; + } + } + } else { + if (verbose) { printf("Error: k %d lower %e f(a_0) %e upper %e\n", k, lower, f_a_0, upper); } + return -1; + } + cut_rhs = (k + 1.0) * std::floor(rhs); + if (verbose) { + printf("Generated strong CG cut: k %d f_a_0 %e cut_rhs %e\n", k, f_a_0, cut_rhs); + for (i_t q = 0; q < cut.i.size(); q++) { + if (cut.x[q] != 0.0) { printf("%.16e x%d ", cut.x[q], cut.i[q]); } + } + printf("\n"); + printf("Original inequality rhs %e nz %ld\n", rhs, coefficients.size()); + for (i_t q = 0; q < nz; q++) { + printf("%e x%d ", coefficients[q], indicies[q]); + } + printf("\n"); + } + return 0; +} + +template +i_t strong_cg_cut_t::generate_strong_cg_cut( + const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + const std::vector& var_types, + const sparse_vector_t& inequality, + const f_t inequality_rhs, + const std::vector& xstar, + sparse_vector_t& cut, + f_t& cut_rhs) +{ +#ifdef PRINT_INEQUALITY_INFO + for (i_t k = 0; k < inequality.i.size(); k++) { + printf("%e %c%d ", + inequality.x[k], + var_types[inequality.i[k]] == variable_type_t::CONTINUOUS ? 'x' : 'y', + inequality.i[k]); + } + printf("CG inequality rhs %e\n", inequality_rhs); +#endif + // Try to remove continuous variables from the inequality + // and transform integer variables to be nonnegative + + // Copy the inequality since remove continuous variables will modify it + sparse_vector_t cg_inequality = inequality; + f_t cg_inequality_rhs = inequality_rhs; + i_t status = remove_continuous_variables_integers_nonnegative( + lp, settings, var_types, cg_inequality, cg_inequality_rhs); + + if (status != 0) { + // Try negating the equality and see if that helps + cg_inequality = inequality; + cg_inequality.negate(); + cg_inequality_rhs = -inequality_rhs; + + status = remove_continuous_variables_integers_nonnegative( + lp, settings, var_types, cg_inequality, cg_inequality_rhs); + } + + if (status == 0) { + // We have an inequality with no continuous variables + + // Generate a CG cut + status = generate_strong_cg_cut_integer_only( + settings, var_types, cg_inequality, cg_inequality_rhs, cut, cut_rhs); + if (status != 0) { return -1; } + + // Convert the CG cut back to the original variables + to_original_integer_variables(lp, cut, cut_rhs); + + // Check for violation + f_t dot = cut.dot(xstar); + // If the cut is violated we will have: sum_j a_j xstar_j > rhs + f_t violation = dot - cut_rhs; + const f_t min_violation_threshold = 1e-6; + if (violation > min_violation_threshold) { + // Note that no slacks are currently present. Since slacks are currently treated as + // continuous. However, this may change. We may need to substitute out the slacks here + + // The CG cut is in the form: sum_j a_j x_j <= rhs + // The cut pool wants the cut in the form: sum_j a_j x_j >= rhs + cut.negate(); + cut_rhs *= -1.0; + return 0; + } + } + return -1; +} + +template +i_t add_cuts(const simplex_solver_settings_t& settings, + const csr_matrix_t& cuts, + const std::vector& cut_rhs, + lp_problem_t& lp, + std::vector& new_slacks, + lp_solution_t& solution, + basis_update_mpf_t& basis_update, + std::vector& basic_list, + std::vector& nonbasic_list, + std::vector& vstatus, + std::vector& edge_norms) + +{ + // Given a set of cuts: C*x <= d that are currently violated + // by the current solution x* (i.e. C*x* > d), this function + // adds the cuts into the LP and solves again. + +#ifdef CHECK_BASIS + { + csc_matrix_t Btest(lp.num_rows, lp.num_rows, 1); + basis_update.multiply_lu(Btest); + csc_matrix_t B(lp.num_rows, lp.num_rows, 1); + form_b(lp.A, basic_list, B); + csc_matrix_t Diff(lp.num_rows, lp.num_rows, 1); + add(Btest, B, 1.0, -1.0, Diff); + const f_t err = Diff.norm1(); + settings.log.printf("Before || B - L*U || %e\n", err); + assert(err <= 1e-6); + } +#endif + + const i_t p = cuts.m; + if (cut_rhs.size() != static_cast(p)) { + settings.log.printf("cut_rhs must have the same number of rows as cuts\n"); + assert(cut_rhs.size() == static_cast(p)); + } + settings.log.debug("Number of cuts %d\n", p); + settings.log.debug("Original lp rows %d\n", lp.num_rows); + settings.log.debug("Original lp cols %d\n", lp.num_cols); + + csr_matrix_t new_A_row(lp.num_rows, lp.num_cols, 1); + lp.A.to_compressed_row(new_A_row); + + i_t append_status = new_A_row.append_rows(cuts); + if (append_status != 0) { + settings.log.printf("append_rows error: %d\n", append_status); + assert(append_status == 0); + } + + csc_matrix_t new_A_col(lp.num_rows + p, lp.num_cols, 1); + new_A_row.to_compressed_col(new_A_col); + + // Add in slacks variables for the new rows + lp.lower.resize(lp.num_cols + p); + lp.upper.resize(lp.num_cols + p); + lp.objective.resize(lp.num_cols + p); + edge_norms.resize(lp.num_cols + p); + i_t nz = new_A_col.col_start[lp.num_cols]; + new_A_col.col_start.resize(lp.num_cols + p + 1); + new_A_col.i.resize(nz + p); + new_A_col.x.resize(nz + p); + i_t k = lp.num_rows; + for (i_t j = lp.num_cols; j < lp.num_cols + p; j++) { + new_A_col.col_start[j] = nz; + new_A_col.i[nz] = k++; + new_A_col.x[nz] = 1.0; + nz++; + lp.lower[j] = 0.0; + lp.upper[j] = inf; + lp.objective[j] = 0.0; + edge_norms[j] = 1.0; + new_slacks.push_back(j); + } + settings.log.debug("Done adding slacks\n"); + new_A_col.col_start[lp.num_cols + p] = nz; + new_A_col.n = lp.num_cols + p; + + lp.A = new_A_col; + + // Check that all slack columns have length 1 + for (i_t slack : new_slacks) { + const i_t col_start = lp.A.col_start[slack]; + const i_t col_end = lp.A.col_start[slack + 1]; + const i_t col_len = col_end - col_start; + if (col_len != 1) { + settings.log.printf("Add cuts: Slack %d has %d nzs in column\n", slack, col_len); + assert(col_len == 1); + } + } + + i_t old_rows = lp.num_rows; + lp.num_rows += p; + i_t old_cols = lp.num_cols; + lp.num_cols += p; + + lp.rhs.resize(lp.num_rows); + for (i_t k = old_rows; k < old_rows + p; k++) { + const i_t h = k - old_rows; + lp.rhs[k] = cut_rhs[h]; + } + settings.log.debug("Done adding rhs\n"); + + // Construct C_B = C(:, basic_list) + std::vector C_col_degree(lp.num_cols, 0); + i_t cuts_nz = cuts.row_start[p]; + for (i_t q = 0; q < cuts_nz; q++) { + const i_t j = cuts.j[q]; + if (j >= lp.num_cols) { + settings.log.printf("Cut column index j=%d exceeds num_cols=%d\n", j, lp.num_cols); + return -1; + } + C_col_degree[j]++; + } + settings.log.debug("Done computing C_col_degree\n"); + + std::vector in_basis(old_cols, -1); + const i_t num_basic = static_cast(basic_list.size()); + i_t C_B_nz = 0; + for (i_t k = 0; k < num_basic; k++) { + const i_t j = basic_list[k]; + if (j < 0 || j >= old_cols) { + settings.log.printf( + "basic_list[%d] = %d is out of bounds %d old_cols %d\n", k, j, j, old_cols); + assert(j >= 0 && j < old_cols); + } + in_basis[j] = k; + // The cuts are on the original variables. So it is possible that + // a slack will be basic and thus not part of the cuts matrix + if (j < cuts.n) { C_B_nz += C_col_degree[j]; } + } + settings.log.debug("Done estimating C_B_nz\n"); + + csr_matrix_t C_B(p, num_basic, C_B_nz); + nz = 0; + for (i_t i = 0; i < p; i++) { + C_B.row_start[i] = nz; + const i_t row_start = cuts.row_start[i]; + const i_t row_end = cuts.row_start[i + 1]; + for (i_t q = row_start; q < row_end; q++) { + const i_t j = cuts.j[q]; + const i_t j_basis = in_basis[j]; + if (j_basis == -1) { continue; } + C_B.j[nz] = j_basis; + C_B.x[nz] = cuts.x[q]; + nz++; + } + } + C_B.row_start[p] = nz; + + if (nz != C_B_nz) { + settings.log.printf("Add cuts: predicted nz %d actual nz %d\n", C_B_nz, nz); + assert(nz == C_B_nz); + } + settings.log.debug("C_B rows %d cols %d nz %d\n", C_B.m, C_B.n, nz); + + // Adjust the basis update to include the new cuts + basis_update.append_cuts(C_B); + + basic_list.resize(lp.num_rows, 0); + i_t h = old_cols; + for (i_t j = old_rows; j < lp.num_rows; j++) { + basic_list[j] = h++; + } + +#ifdef CHECK_BASIS + // Check the basis update + csc_matrix_t Btest(lp.num_rows, lp.num_rows, 1); + basis_update.multiply_lu(Btest); + + csc_matrix_t B(lp.num_rows, lp.num_rows, 1); + form_b(lp.A, basic_list, B); + + csc_matrix_t Diff(lp.num_rows, lp.num_rows, 1); + add(Btest, B, 1.0, -1.0, Diff); + const f_t err = Diff.norm1(); + settings.log.printf("After || B - L*U || %e\n", err); + if (err > 1e-6) { + settings.log.printf("Diff matrix\n"); + // Diff.print_matrix(); + assert(err <= 1e-6); + } +#endif + // Adjust the vstatus + vstatus.resize(lp.num_cols); + for (i_t j = old_cols; j < lp.num_cols; j++) { + vstatus[j] = variable_status_t::BASIC; + } + + // Adjust the solution + solution.x.resize(lp.num_cols, 0.0); + solution.y.resize(lp.num_rows, 0.0); + solution.z.resize(lp.num_cols, 0.0); + + return 0; +} + +template +void remove_cuts(lp_problem_t& lp, + const simplex_solver_settings_t& settings, + csr_matrix_t& Arow, + std::vector& new_slacks, + i_t original_rows, + std::vector& var_types, + std::vector& vstatus, + std::vector& edge_norms, + std::vector& x, + std::vector& y, + std::vector& z, + std::vector& basic_list, + std::vector& nonbasic_list, + basis_update_mpf_t& basis_update) +{ + std::vector cuts_to_remove; + cuts_to_remove.reserve(lp.num_rows - original_rows); + std::vector slacks_to_remove; + slacks_to_remove.reserve(lp.num_rows - original_rows); + const f_t dual_tol = 1e-10; + + std::vector is_slack(lp.num_cols, 0); + for (i_t j : new_slacks) { + is_slack[j] = 1; +#ifdef CHECK_SLACKS + // Check that slack column length is 1 + const i_t col_start = lp.A.col_start[j]; + const i_t col_end = lp.A.col_start[j + 1]; + const i_t col_len = col_end - col_start; + if (col_len != 1) { + printf("Remove cuts: Slack %d has %d nzs in column\n", j, col_len); + assert(col_len == 1); + } +#endif + } + + for (i_t k = original_rows; k < lp.num_rows; k++) { + if (std::abs(y[k]) < dual_tol) { + const i_t row_start = Arow.row_start[k]; + const i_t row_end = Arow.row_start[k + 1]; + i_t last_slack = -1; + const f_t slack_tol = 1e-3; + for (i_t p = row_start; p < row_end; p++) { + const i_t j = Arow.j[p]; + if (is_slack[j]) { + if (vstatus[j] == variable_status_t::BASIC && x[j] > slack_tol) { last_slack = j; } + } + } + if (last_slack != -1) { + cuts_to_remove.push_back(k); + slacks_to_remove.push_back(last_slack); + } + } + } + + if (cuts_to_remove.size() > 0) { + std::vector marked_rows(lp.num_rows, 0); + for (i_t i : cuts_to_remove) { + marked_rows[i] = 1; + } + std::vector marked_cols(lp.num_cols, 0); + for (i_t j : slacks_to_remove) { + marked_cols[j] = 1; + } + + std::vector new_rhs(lp.num_rows - cuts_to_remove.size()); + std::vector new_solution_y(lp.num_rows - cuts_to_remove.size()); + i_t h = 0; + for (i_t i = 0; i < lp.num_rows; i++) { + if (!marked_rows[i]) { + new_rhs[h] = lp.rhs[i]; + new_solution_y[h] = y[i]; + h++; + } + } + csr_matrix_t new_Arow(1, 1, 0); + Arow.remove_rows(marked_rows, new_Arow); + Arow = new_Arow; + Arow.to_compressed_col(lp.A); + + std::vector new_objective(lp.num_cols - slacks_to_remove.size()); + std::vector new_lower(lp.num_cols - slacks_to_remove.size()); + std::vector new_upper(lp.num_cols - slacks_to_remove.size()); + std::vector new_var_types(lp.num_cols - slacks_to_remove.size()); + std::vector new_vstatus(lp.num_cols - slacks_to_remove.size()); + std::vector new_edge_norms(lp.num_cols - slacks_to_remove.size()); + std::vector new_basic_list; + new_basic_list.reserve(lp.num_rows - slacks_to_remove.size()); + std::vector new_nonbasic_list; + new_nonbasic_list.reserve(nonbasic_list.size()); + std::vector new_solution_x(lp.num_cols - slacks_to_remove.size()); + std::vector new_solution_z(lp.num_cols - slacks_to_remove.size()); + std::vector new_is_slacks(lp.num_cols - slacks_to_remove.size(), 0); + h = 0; + for (i_t k = 0; k < lp.num_cols; k++) { + if (!marked_cols[k]) { + new_objective[h] = lp.objective[k]; + new_lower[h] = lp.lower[k]; + new_upper[h] = lp.upper[k]; + new_var_types[h] = var_types[k]; + new_vstatus[h] = vstatus[k]; + new_edge_norms[h] = edge_norms[k]; + new_solution_x[h] = x[k]; + new_solution_z[h] = z[k]; + new_is_slacks[h] = is_slack[k]; + if (new_vstatus[h] != variable_status_t::BASIC) { + new_nonbasic_list.push_back(h); + } else { + new_basic_list.push_back(h); + } + h++; + } + } + lp.A.remove_columns(marked_cols); + lp.A.to_compressed_row(Arow); + lp.objective = new_objective; + lp.lower = new_lower; + lp.upper = new_upper; + lp.rhs = new_rhs; + var_types = new_var_types; + lp.num_cols = lp.A.n; + lp.num_rows = lp.A.m; + + new_slacks.clear(); + new_slacks.reserve(lp.num_cols); + for (i_t j = 0; j < lp.num_cols; j++) { + if (new_is_slacks[j]) { new_slacks.push_back(j); } + } + basic_list = new_basic_list; + nonbasic_list = new_nonbasic_list; + vstatus = new_vstatus; + edge_norms = new_edge_norms; + x = new_solution_x; + y = new_solution_y; + z = new_solution_z; + + settings.log.debug("Removed %d cuts. After removal %d rows %d columns %d nonzeros\n", + cuts_to_remove.size(), + lp.num_rows, + lp.num_cols, + lp.A.col_start[lp.A.n]); + + basis_update.resize(lp.num_rows); + basis_update.refactor_basis( + lp.A, settings, lp.lower, lp.upper, basic_list, nonbasic_list, vstatus); + } +} + +template +void read_saved_solution_for_cut_verification(const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + std::vector& saved_solution) +{ + settings.log.printf("Trying to open solution.dat\n"); + FILE* fid = NULL; + fid = fopen("solution.dat", "r"); + if (fid != NULL) { + i_t n_solution_dat; + i_t count = fscanf(fid, "%d\n", &n_solution_dat); + settings.log.printf( + "Solution.dat variables %d =? %d =? count %d\n", n_solution_dat, lp.num_cols, count); + bool good = true; + if (count == 1 && n_solution_dat == lp.num_cols) { + settings.log.printf("Opened solution.dat with %d number of variables\n", n_solution_dat); + saved_solution.resize(n_solution_dat); + for (i_t j = 0; j < n_solution_dat; j++) { + count = fscanf(fid, "%lf", &saved_solution[j]); + if (count != 1) { + settings.log.printf("bad read solution.dat: j %d count %d\n", j, count); + good = false; + break; + } + } + } else { + good = false; + } + fclose(fid); + + if (!good) { + saved_solution.resize(0); + settings.log.printf("Solution.dat is bad\n"); + } else { + settings.log.printf("Read solution file\n"); + + auto hash_combine_f = [](size_t seed, f_t x) { + seed ^= std::hash{}(x) + 0x9e3779b9 + (seed << 6) + (seed >> 2); + return seed; + }; + size_t seed = lp.num_cols; + for (i_t j = 0; j < lp.num_cols; ++j) { + seed = hash_combine_f(seed, saved_solution[j]); + } + settings.log.printf("Saved solution hash: %20x\n", seed); + + // Compute || A * x - b ||_inf + std::vector residual = lp.rhs; + matrix_vector_multiply(lp.A, 1.0, saved_solution, -1.0, residual); + settings.log.printf("Saved solution: || A*x - b ||_inf %e\n", + vector_norm_inf(residual)); + f_t infeas = 0; + for (i_t j = 0; j < lp.num_cols; j++) { + if (saved_solution[j] < lp.lower[j] - 1e-6) { + f_t curr_infeas = (lp.lower[j] - saved_solution[j]); + infeas += curr_infeas; + settings.log.printf( + "j: %d saved solution %e lower %e\n", j, saved_solution[j], lp.lower[j]); + } + if (saved_solution[j] > lp.upper[j] + 1e-6) { + f_t curr_infeas = (saved_solution[j] - lp.upper[j]); + infeas += curr_infeas; + settings.log.printf( + "j %d saved solution %e upper %e\n", j, saved_solution[j], lp.upper[j]); + } + } + settings.log.printf("Bound infeasibility %e\n", infeas); + } + } else { + settings.log.printf("Could not open solution.dat\n"); + } +} + +template +void write_solution_for_cut_verification(const lp_problem_t& lp, + const std::vector& solution) +{ + FILE* fid = NULL; + fid = fopen("solution.dat", "w"); + if (fid != NULL) { + printf("Writing solution.dat\n"); + + std::vector residual = lp.rhs; + matrix_vector_multiply(lp.A, 1.0, solution, -1.0, residual); + printf("|| A*x - b ||_inf %e\n", vector_norm_inf(residual)); + auto hash_combine_f = [](size_t seed, f_t x) { + seed ^= std::hash{}(x) + 0x9e3779b9 + (seed << 6) + (seed >> 2); + return seed; + }; + printf("incumbent size %ld original lp cols %d\n", solution.size(), lp.num_cols); + i_t n = lp.num_cols; + size_t seed = n; + fprintf(fid, "%d\n", n); + for (i_t j = 0; j < n; ++j) { + fprintf(fid, "%.17g\n", solution[j]); + seed = hash_combine_f(seed, solution[j]); + } + printf("Solution hash: %20x\n", seed); + fclose(fid); + } +} + +template +void verify_cuts_against_saved_solution(const csr_matrix_t& cuts, + const std::vector& cut_rhs, + const std::vector& saved_solution) +{ + if (saved_solution.size() > 0) { + csc_matrix_t cuts_to_add_col(cuts.m, cuts.n, cuts.row_start[cuts.m]); + cuts.to_compressed_col(cuts_to_add_col); + std::vector Cx(cuts.m); + matrix_vector_multiply(cuts_to_add_col, 1.0, saved_solution, 0.0, Cx); + const i_t num_cuts = cuts.m; + for (i_t k = 0; k < num_cuts; k++) { + if (Cx[k] > cut_rhs[k] + 1e-6) { + printf("Cut %d is violated by saved solution. Cx %e cut_rhs %e Diff: %e\n", + k, + Cx[k], + cut_rhs[k], + Cx[k] - cut_rhs[k]); + } + } + } +} + +#ifdef DUAL_SIMPLEX_INSTANTIATE_DOUBLE +template class cut_pool_t; +template class cut_generation_t; +template class tableau_equality_t; +template class mixed_integer_rounding_cut_t; + +template int add_cuts(const simplex_solver_settings_t& settings, + const csr_matrix_t& cuts, + const std::vector& cut_rhs, + lp_problem_t& lp, + std::vector& new_slacks, + lp_solution_t& solution, + basis_update_mpf_t& basis_update, + std::vector& basic_list, + std::vector& nonbasic_list, + std::vector& vstatus, + std::vector& edge_norms); + +template void remove_cuts(lp_problem_t& lp, + const simplex_solver_settings_t& settings, + csr_matrix_t& Arow, + std::vector& new_slacks, + int original_rows, + std::vector& var_types, + std::vector& vstatus, + std::vector& edge_norms, + std::vector& x, + std::vector& y, + std::vector& z, + std::vector& basic_list, + std::vector& nonbasic_list, + basis_update_mpf_t& basis_update); + +template void read_saved_solution_for_cut_verification( + const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + std::vector& saved_solution); + +template void write_solution_for_cut_verification(const lp_problem_t& lp, + const std::vector& solution); + +template void verify_cuts_against_saved_solution( + const csr_matrix_t& cuts, + const std::vector& cut_rhs, + const std::vector& saved_solution); +#endif + +} // namespace cuopt::linear_programming::dual_simplex diff --git a/cpp/src/dual_simplex/cuts.hpp b/cpp/src/dual_simplex/cuts.hpp new file mode 100644 index 0000000000..a4a36d75b2 --- /dev/null +++ b/cpp/src/dual_simplex/cuts.hpp @@ -0,0 +1,479 @@ +/* clang-format off */ +/* + * SPDX-FileCopyrightText: Copyright (c) 2025-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. + * SPDX-License-Identifier: Apache-2.0 + */ +/* clang-format on */ +#pragma once + +#include +#include +#include +#include +#include +#include + +#include +#include +#include +#include +#include + +#include +#include + +namespace cuopt::linear_programming::dual_simplex { + +enum cut_type_t : int8_t { + MIXED_INTEGER_GOMORY = 0, + MIXED_INTEGER_ROUNDING = 1, + KNAPSACK = 2, + CHVATAL_GOMORY = 3, + MAX_CUT_TYPE = 4 +}; + +template +struct cut_info_t { + bool has_cuts() const + { + i_t total_cuts = 0; + for (i_t i = 0; i < MAX_CUT_TYPE; i++) { + total_cuts += num_cuts[i]; + } + return total_cuts > 0; + } + void record_cut_types(const std::vector& cut_types) + { + for (cut_type_t cut_type : cut_types) { + num_cuts[static_cast(cut_type)]++; + } + } + const char* cut_type_names[MAX_CUT_TYPE] = {"Gomory ", "MIR ", "Knapsack ", "Strong CG"}; + std::array num_cuts = {0}; +}; + +template +void print_cut_info(const simplex_solver_settings_t& settings, + const cut_info_t& cut_info) +{ + if (cut_info.has_cuts()) { + for (i_t i = 0; i < MAX_CUT_TYPE; i++) { + settings.log.printf("%s cuts : %d\n", cut_info.cut_type_names[i], cut_info.num_cuts[i]); + } + } +} + +template +void print_cut_types(const std::string& prefix, + const std::vector& cut_types, + const simplex_solver_settings_t& settings) +{ + cut_info_t cut_info; + cut_info.record_cut_types(cut_types); + settings.log.printf("%s: ", prefix.c_str()); + for (i_t i = 0; i < MAX_CUT_TYPE; i++) { + settings.log.printf("%s cuts: %d ", cut_info.cut_type_names[i], cut_info.num_cuts[i]); + if (i < MAX_CUT_TYPE - 1) { settings.log.printf(", "); } + } + settings.log.printf("\n"); +} + +template +f_t fractional_part(f_t a) +{ + return a - std::floor(a); +} + +// Computes a permutation of a score vector that puts the highest scores first +template +void best_score_first_permutation(std::vector& scores, std::vector& permutation) +{ + if (permutation.size() != scores.size()) { permutation.resize(scores.size()); } + std::iota(permutation.begin(), permutation.end(), 0); + std::sort( + permutation.begin(), permutation.end(), [&](i_t a, i_t b) { return scores[a] > scores[b]; }); +} + +// Computes a permutation of a score vector that puts the highest score last +template +void best_score_last_permutation(std::vector& scores, std::vector& permutation) +{ + if (permutation.size() != scores.size()) { permutation.resize(scores.size()); } + std::iota(permutation.begin(), permutation.end(), 0); + std::sort( + permutation.begin(), permutation.end(), [&](i_t a, i_t b) { return scores[a] < scores[b]; }); +} + +// Routines for verifying cuts against a saved solution +template +void read_saved_solution_for_cut_verification(const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + std::vector& saved_solution); + +template +void write_solution_for_cut_verification(const lp_problem_t& lp, + const std::vector& solution); + +template +void verify_cuts_against_saved_solution(const csr_matrix_t& cuts, + const std::vector& cut_rhs, + const std::vector& saved_solution); + +template +class cut_pool_t { + public: + cut_pool_t(i_t original_vars, const simplex_solver_settings_t& settings) + : original_vars_(original_vars), + settings_(settings), + cut_storage_(0, original_vars, 0), + rhs_storage_(0), + cut_age_(0), + cut_type_(0), + scored_cuts_(0) + { + } + + // Add a cut in the form: cut'*x >= rhs. + // We expect that the cut is violated by the current relaxation xstar + // cut'*xstart < rhs + void add_cut(cut_type_t cut_type, const sparse_vector_t& cut, f_t rhs); + + void score_cuts(std::vector& x_relax); + + // We return the cuts in the form best_cuts*x <= best_rhs + i_t get_best_cuts(csr_matrix_t& best_cuts, + std::vector& best_rhs, + std::vector& best_cut_types); + + void age_cuts(); + + void drop_cuts(); + + i_t pool_size() const { return cut_storage_.m; } + + void print_cutpool_types() { print_cut_types("In cut pool", cut_type_, settings_); } + + private: + f_t cut_distance(i_t row, const std::vector& x, f_t& cut_violation, f_t& cut_norm); + f_t cut_density(i_t row); + f_t cut_orthogonality(i_t i, i_t j); + + i_t original_vars_; + const simplex_solver_settings_t& settings_; + + csr_matrix_t cut_storage_; + std::vector rhs_storage_; + std::vector cut_age_; + std::vector cut_type_; + + i_t scored_cuts_; + std::vector cut_distances_; + std::vector cut_norms_; + std::vector cut_orthogonality_; + std::vector cut_scores_; + std::vector best_cuts_; +}; + +template +class knapsack_generation_t { + public: + knapsack_generation_t(const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + csr_matrix_t& Arow, + const std::vector& new_slacks, + const std::vector& var_types); + + i_t generate_knapsack_cuts(const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + csr_matrix_t& Arow, + const std::vector& new_slacks, + const std::vector& var_types, + const std::vector& xstar, + i_t knapsack_row, + sparse_vector_t& cut, + f_t& cut_rhs); + + i_t num_knapsack_constraints() const { return knapsack_constraints_.size(); } + const std::vector& get_knapsack_constraints() const { return knapsack_constraints_; } + + private: + // Generate a heuristic solution to the 0-1 knapsack problem + f_t greedy_knapsack_problem(const std::vector& values, + const std::vector& weights, + f_t rhs, + std::vector& solution); + + // Solve a 0-1 knapsack problem using dynamic programming + f_t solve_knapsack_problem(const std::vector& values, + const std::vector& weights, + f_t rhs, + std::vector& solution); + + std::vector is_slack_; + std::vector knapsack_constraints_; + const simplex_solver_settings_t& settings_; +}; + +// Forward declaration +template +class mixed_integer_rounding_cut_t; + +template +class cut_generation_t { + public: + cut_generation_t(cut_pool_t& cut_pool, + const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + csr_matrix_t& Arow, + const std::vector& new_slacks, + const std::vector& var_types) + : cut_pool_(cut_pool), knapsack_generation_(lp, settings, Arow, new_slacks, var_types) + { + } + + void generate_cuts(const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + csr_matrix_t& Arow, + const std::vector& new_slacks, + const std::vector& var_types, + basis_update_mpf_t& basis_update, + const std::vector& xstar, + const std::vector& basic_list, + const std::vector& nonbasic_list); + + private: + // Generate all mixed integer gomory cuts + void generate_gomory_cuts(const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + csr_matrix_t& Arow, + const std::vector& new_slacks, + const std::vector& var_types, + basis_update_mpf_t& basis_update, + const std::vector& xstar, + const std::vector& basic_list, + const std::vector& nonbasic_list); + + // Generate all mixed integer rounding cuts + void generate_mir_cuts(const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + csr_matrix_t& Arow, + const std::vector& new_slacks, + const std::vector& var_types, + const std::vector& xstar); + + // Generate all knapsack cuts + void generate_knapsack_cuts(const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + csr_matrix_t& Arow, + const std::vector& new_slacks, + const std::vector& var_types, + const std::vector& xstar); + + cut_pool_t& cut_pool_; + knapsack_generation_t knapsack_generation_; +}; + +template +class tableau_equality_t { + public: + tableau_equality_t(const lp_problem_t& lp, + basis_update_mpf_t& basis_update, + const std::vector& nonbasic_list) + : b_bar_(lp.num_rows, 0.0), + nonbasic_mark_(lp.num_cols, 0), + x_workspace_(lp.num_cols, 0.0), + x_mark_(lp.num_cols, 0), + c_workspace_(lp.num_cols, 0.0) + { + basis_update.b_solve(lp.rhs, b_bar_); + for (i_t j : nonbasic_list) { + nonbasic_mark_[j] = 1; + } + } + + // Generates the base inequalities: C*x == d that will be turned into cuts + i_t generate_base_equality(const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + csr_matrix_t& Arow, + const std::vector& var_types, + basis_update_mpf_t& basis_update, + const std::vector& xstar, + const std::vector& basic_list, + const std::vector& nonbasic_list, + i_t i, + sparse_vector_t& inequality, + f_t& inequality_rhs); + + private: + std::vector b_bar_; + std::vector nonbasic_mark_; + std::vector x_workspace_; + std::vector x_mark_; + std::vector c_workspace_; +}; + +template +class mixed_integer_rounding_cut_t { + public: + mixed_integer_rounding_cut_t(const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + const std::vector& new_slacks, + const std::vector& xstar); + + // Convert an inequality of the form: sum_j a_j x_j >= beta + // with l_j <= x_j <= u_j into the form: + // sum_{j not in L union U} d_j x_j + sum_{j in L} d_j v_j + // + sum_{j in U} d_j w_j >= delta, + // where v_j = x_j - l_j for j in L + // and w_j = u_j - x_j for j in Us + void to_nonnegative(const lp_problem_t& lp, + sparse_vector_t& inequality, + f_t& rhs); + + void relaxation_to_nonnegative(const lp_problem_t& lp, + const std::vector& xstar, + std::vector& xstar_nonnegative); + + // Convert an inequality of the form: + // sum_{j not in L union U} d_j x_j + sum_{j in L} d_j v_j + // + sum_{j in U} d_j w_j >= delta + // where v_j = x_j - l_j for j in L + // and w_j = u_j - x_j for j in U + // back to an inequality on the original variables + // sum_j a_j x_j >= beta + void to_original(const lp_problem_t& lp, + sparse_vector_t& inequality, + f_t& rhs); + + // Given a cut of the form sum_j d_j x_j >= beta + // with l_j <= x_j <= u_j, try to remove coefficients d_j + // with | d_j | < epsilon + void remove_small_coefficients(const std::vector& lower_bounds, + const std::vector& upper_bounds, + sparse_vector_t& cut, + f_t& cut_rhs); + + // Given an inequality sum_j a_j x_j >= beta, x_j >= 0, x_j in Z, j in I + // generate an MIR cut of the form sum_j d_j x_j >= delta + i_t generate_cut_nonnegative(const sparse_vector_t& a, + f_t beta, + const std::vector& var_types, + sparse_vector_t& cut, + f_t& cut_rhs); + + f_t compute_violation(const sparse_vector_t& cut, + f_t cut_rhs, + const std::vector& xstar); + + i_t generate_cut(const sparse_vector_t& a, + f_t beta, + const std::vector& upper_bounds, + const std::vector& lower_bounds, + const std::vector& var_types, + sparse_vector_t& cut, + f_t& cut_rhs); + + void substitute_slacks(const lp_problem_t& lp, + csr_matrix_t& Arow, + sparse_vector_t& cut, + f_t& cut_rhs); + + // Combine the pivot row with the inequality to eliminate the variable j + // The new inequality is returned in inequality and inequality_rhs + void combine_rows(const lp_problem_t& lp, + csr_matrix_t& Arow, + i_t j, + const sparse_vector_t& pivot_row, + f_t pivot_row_rhs, + sparse_vector_t& inequality, + f_t& inequality_rhs); + + private: + i_t num_vars_; + const simplex_solver_settings_t& settings_; + std::vector x_workspace_; + std::vector x_mark_; + std::vector has_lower_; + std::vector has_upper_; + std::vector is_slack_; + std::vector slack_rows_; + std::vector indices_; + std::vector bound_info_; + bool needs_complement_; +}; + +template +class strong_cg_cut_t { + public: + strong_cg_cut_t(const lp_problem_t& lp, + const std::vector& var_types, + const std::vector& xstar); + + i_t generate_strong_cg_cut(const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + const std::vector& var_types, + const sparse_vector_t& inequality, + const f_t inequality_rhs, + const std::vector& xstar, + sparse_vector_t& cut, + f_t& cut_rhs); + + i_t remove_continuous_variables_integers_nonnegative( + const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + const std::vector& var_types, + sparse_vector_t& inequality, + f_t& inequality_rhs); + + void to_original_integer_variables(const lp_problem_t& lp, + sparse_vector_t& cut, + f_t& cut_rhs); + + i_t generate_strong_cg_cut_integer_only(const simplex_solver_settings_t& settings, + const std::vector& var_types, + const sparse_vector_t& inequality, + f_t inequality_rhs, + sparse_vector_t& cut, + f_t& cut_rhs); + + private: + i_t generate_strong_cg_cut_helper(const std::vector& indicies, + const std::vector& coefficients, + f_t rhs, + const std::vector& var_types, + sparse_vector_t& cut, + f_t& cut_rhs); + + std::vector transformed_variables_; +}; + +template +i_t add_cuts(const simplex_solver_settings_t& settings, + const csr_matrix_t& cuts, + const std::vector& cut_rhs, + lp_problem_t& lp, + std::vector& new_slacks, + lp_solution_t& solution, + basis_update_mpf_t& basis_update, + std::vector& basic_list, + std::vector& nonbasic_list, + std::vector& vstatus, + std::vector& edge_norms); + +template +void remove_cuts(lp_problem_t& lp, + const simplex_solver_settings_t& settings, + csr_matrix_t& Arow, + std::vector& new_slacks, + i_t original_rows, + std::vector& var_types, + std::vector& vstatus, + std::vector& edge_norms, + std::vector& x, + std::vector& y, + std::vector& z, + std::vector& basic_list, + std::vector& nonbasic_list, + basis_update_mpf_t& basis_update); + +} // namespace cuopt::linear_programming::dual_simplex diff --git a/cpp/src/dual_simplex/dense_matrix.hpp b/cpp/src/dual_simplex/dense_matrix.hpp index b1fc521b32..7f3f08813c 100644 --- a/cpp/src/dual_simplex/dense_matrix.hpp +++ b/cpp/src/dual_simplex/dense_matrix.hpp @@ -1,6 +1,6 @@ /* clang-format off */ /* - * SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved. + * SPDX-FileCopyrightText: Copyright (c) 2025-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. * SPDX-License-Identifier: Apache-2.0 */ /* clang-format on */ @@ -18,6 +18,8 @@ class dense_matrix_t { public: dense_matrix_t(i_t rows, i_t cols) : m(rows), n(cols), values(rows * cols, 0.0) {} + dense_matrix_t(i_t rows, i_t cols, f_t value) : m(rows), n(cols), values(rows * cols, value) {} + void resize(i_t rows, i_t cols) { m = rows; diff --git a/cpp/src/dual_simplex/folding.cpp b/cpp/src/dual_simplex/folding.cpp index 6eae5eb1ed..c59d827c51 100644 --- a/cpp/src/dual_simplex/folding.cpp +++ b/cpp/src/dual_simplex/folding.cpp @@ -1,6 +1,6 @@ /* clang-format off */ /* - * SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved. + * SPDX-FileCopyrightText: Copyright (c) 2025-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. * SPDX-License-Identifier: Apache-2.0 */ /* clang-format on */ @@ -386,16 +386,18 @@ void color_lower_bounds(const csc_matrix_t& A, col_lower_bound = static_cast(unique_col_sums.size()); } +enum coloring_status_t : int8_t { COLORING_SUCCESS = 0, COLORING_FAILED = -1 }; + template -i_t color_graph(const csc_matrix_t& A, - const simplex_solver_settings_t& settings, - std::vector>& colors, - i_t row_threshold, - i_t col_threshold, - i_t& num_row_colors, - i_t& num_col_colors, - i_t& num_colors, - i_t& total_colors_seen) +coloring_status_t color_graph(const csc_matrix_t& A, + const simplex_solver_settings_t& settings, + std::vector>& colors, + i_t row_threshold, + i_t col_threshold, + i_t& num_row_colors, + i_t& num_col_colors, + i_t& num_colors, + i_t& total_colors_seen) { f_t start_time = tic(); f_t last_log_time = start_time; @@ -421,7 +423,7 @@ i_t color_graph(const csc_matrix_t& A, row_threshold, col_lower_bound, col_threshold); - return -1; + return coloring_status_t::COLORING_FAILED; } std::vector all_rows_vertices(m); @@ -577,7 +579,7 @@ i_t color_graph(const csc_matrix_t& A, for (i_t k = 0; k < max_vertices; k++) { if (vertex_to_sum[k] != 0.0) { settings.log.printf("Folding: Vertex %d has sum %e\n", k, vertex_to_sum[k]); - return -2; + return coloring_status_t::COLORING_FAILED; } } #endif @@ -594,7 +596,7 @@ i_t color_graph(const csc_matrix_t& A, settings.log.printf("Folding: Color %d has %ld vertices to refine. Not cleared\n", k, vertices_to_refine_by_color[k].size()); - return -2; + return coloring_status_t::COLORING_FAILED; } } #endif @@ -604,7 +606,7 @@ i_t color_graph(const csc_matrix_t& A, if (row_color_map[i] >= total_colors_seen) { settings.log.printf("Folding: Row color %d is not in the colors vector\n", row_color_map[i]); - return -2; + return coloring_status_t::COLORING_FAILED; } } for (i_t j = 0; j < n; j++) { @@ -612,7 +614,7 @@ i_t color_graph(const csc_matrix_t& A, settings.log.printf("Folding: Column color %d is not in the colors vector. %d\n", col_color_map[j], num_colors); - return -2; + return coloring_status_t::COLORING_FAILED; } } #endif @@ -634,7 +636,7 @@ i_t color_graph(const csc_matrix_t& A, row_color_map[v], color.color, v); - return -2; + return coloring_status_t::COLORING_FAILED; } } } else { @@ -646,7 +648,7 @@ i_t color_graph(const csc_matrix_t& A, col_color_map[v], color.color, v); - return -2; + return coloring_status_t::COLORING_FAILED; } } } @@ -655,19 +657,19 @@ i_t color_graph(const csc_matrix_t& A, // printf("Number of active colors: %d\n", num_active_colors); if (num_active_colors != num_colors) { settings.log.printf("Folding: Number of active colors does not match number of colors\n"); - return -2; + return coloring_status_t::COLORING_FAILED; } // printf("Number of active row colors: %d\n", num_active_row_colors); if (num_active_row_colors != num_row_colors) { settings.log.printf( "Folding: Number of active row colors does not match number of row colors\n"); - return -2; + return coloring_status_t::COLORING_FAILED; } // printf("Number of active column colors: %d\n", num_active_col_colors); if (num_active_col_colors != num_col_colors) { settings.log.printf( "Folding: Number of active column colors does not match number of column colors\n"); - return -2; + return coloring_status_t::COLORING_FAILED; } #endif @@ -694,23 +696,23 @@ i_t color_graph(const csc_matrix_t& A, } if (num_row_colors >= max_vertices) { settings.log.printf("Folding: Too many row colors %d max %d\n", num_row_colors, max_vertices); - return -2; + return coloring_status_t::COLORING_FAILED; } if (num_col_colors >= max_vertices) { settings.log.printf( "Folding: Too many column colors %d max %d\n", num_col_colors, max_vertices); - return -2; + return coloring_status_t::COLORING_FAILED; } if (num_row_colors > row_threshold || num_col_colors > col_threshold) { settings.log.printf("Folding: Number of colors exceeds threshold"); - return -1; + return coloring_status_t::COLORING_FAILED; } } settings.log.printf( "Folding: Colors %d. Refinements: %d\n", num_row_colors + num_col_colors, num_refinements); - return 0; + return coloring_status_t::COLORING_SUCCESS; } template @@ -870,16 +872,17 @@ void folding(lp_problem_t& problem, f_t fold_threshold = settings.folding == -1 ? 0.50 : 1.0; i_t row_threshold = static_cast(fold_threshold * static_cast(m)); i_t col_threshold = static_cast(fold_threshold * static_cast(n)); - i_t status = color_graph(augmented, - settings, - colors, - row_threshold, - col_threshold, - num_row_colors, - num_col_colors, - num_colors, - total_colors_seen); - if (status != 0) { + + coloring_status_t status = color_graph(augmented, + settings, + colors, + row_threshold, + col_threshold, + num_row_colors, + num_col_colors, + num_colors, + total_colors_seen); + if (status != coloring_status_t::COLORING_SUCCESS) { settings.log.printf("Folding: Coloring aborted in %.2f seconds\n", toc(color_start_time)); return; } diff --git a/cpp/src/dual_simplex/mip_node.hpp b/cpp/src/dual_simplex/mip_node.hpp index de147132a3..5ee4f49d13 100644 --- a/cpp/src/dual_simplex/mip_node.hpp +++ b/cpp/src/dual_simplex/mip_node.hpp @@ -60,6 +60,7 @@ class mip_node_t { node_id(0), branch_var(-1), branch_dir(rounding_direction_t::NONE), + integer_infeasible(-1), objective_estimate(std::numeric_limits::infinity()), vstatus(basis) { @@ -73,6 +74,7 @@ class mip_node_t { i_t branch_variable, rounding_direction_t branch_direction, f_t branch_var_value, + i_t integer_inf, const std::vector& basis) : status(node_status_t::PENDING), lower_bound(parent_node->lower_bound), @@ -82,9 +84,9 @@ class mip_node_t { branch_var(branch_variable), branch_dir(branch_direction), fractional_val(branch_var_value), + integer_infeasible(integer_inf), objective_estimate(parent_node->objective_estimate), vstatus(basis) - { branch_var_lower = branch_direction == rounding_direction_t::DOWN ? problem.lower[branch_var] : std::ceil(branch_var_value); @@ -250,6 +252,7 @@ class mip_node_t { f_t branch_var_lower; f_t branch_var_upper; f_t fractional_val; + i_t integer_infeasible; mip_node_t* parent; std::unique_ptr children[2]; @@ -285,6 +288,7 @@ class search_tree_t { void branch(mip_node_t* parent_node, const i_t branch_var, const f_t fractional_val, + const i_t integer_infeasible, const std::vector& parent_vstatus, const lp_problem_t& original_lp, logger_t& log) @@ -297,8 +301,8 @@ class search_tree_t { branch_var, rounding_direction_t::DOWN, fractional_val, + integer_infeasible, parent_vstatus); - graphviz_edge(log, parent_node, down_child.get(), @@ -312,6 +316,7 @@ class search_tree_t { branch_var, rounding_direction_t::UP, fractional_val, + integer_infeasible, parent_vstatus); graphviz_edge(log, diff --git a/cpp/src/dual_simplex/phase2.cpp b/cpp/src/dual_simplex/phase2.cpp index a54101ec88..c74841c1cf 100644 --- a/cpp/src/dual_simplex/phase2.cpp +++ b/cpp/src/dual_simplex/phase2.cpp @@ -1232,7 +1232,9 @@ i_t initialize_steepest_edge_norms(const lp_problem_t& lp, settings.log.printf("Initialized %d of %d steepest edge norms in %.2fs\n", k, m, now); } if (toc(start_time) > settings.time_limit) { return -1; } - if (settings.concurrent_halt != nullptr && *settings.concurrent_halt == 1) { return -1; } + if (settings.concurrent_halt != nullptr && *settings.concurrent_halt == 1) { + return CONCURRENT_HALT_RETURN; + } } return 0; } @@ -1733,6 +1735,69 @@ f_t dual_infeasibility(const lp_problem_t& lp, return sum_infeasible; } +template +f_t primal_infeasibility_breakdown(const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + const std::vector& vstatus, + const std::vector& x, + f_t& basic_infeas, + f_t& nonbasic_infeas, + f_t& basic_over) +{ + const i_t n = lp.num_cols; + f_t primal_inf = 0; + basic_infeas = 0.0; + basic_over = 0.0; + nonbasic_infeas = 0.0; + for (i_t j = 0; j < n; ++j) { + if (x[j] < lp.lower[j]) { + // x_j < l_j => -x_j > -l_j => -x_j + l_j > 0 + const f_t infeas = -x[j] + lp.lower[j]; + if (vstatus[j] == variable_status_t::BASIC) { + basic_infeas += infeas; + if (infeas > settings.primal_tol) { basic_over += infeas; } + } else { + nonbasic_infeas += infeas; + } + primal_inf += infeas; +#ifdef PRIMAL_INFEASIBLE_DEBUG + if (infeas > settings.primal_tol) { + settings.log.printf("x %d infeas %e lo %e val %e up %e vstatus %d\n", + j, + infeas, + lp.lower[j], + x[j], + lp.upper[j], + static_cast(vstatus[j])); + } +#endif + } + if (x[j] > lp.upper[j]) { + // x_j > u_j => x_j - u_j > 0 + const f_t infeas = x[j] - lp.upper[j]; + if (vstatus[j] == variable_status_t::BASIC) { + basic_infeas += infeas; + if (infeas > settings.primal_tol) { basic_over += infeas; } + } else { + nonbasic_infeas += infeas; + } + primal_inf += infeas; +#ifdef PRIMAL_INFEASIBLE_DEBUG + if (infeas > settings.primal_tol) { + settings.log.printf("x %d infeas %e lo %e val %e up %e vstatus %d\n", + j, + infeas, + lp.lower[j], + x[j], + lp.upper[j], + static_cast(vstatus[j])); + } +#endif + } + } + return primal_inf; +} + template f_t primal_infeasibility(const lp_problem_t& lp, const simplex_solver_settings_t& settings, @@ -1914,6 +1979,7 @@ void set_primal_variables_on_bounds(const lp_problem_t& lp, std::vector& x) { const i_t n = lp.num_cols; + f_t tol = 1e-10; for (i_t j = 0; j < n; ++j) { // We set z_j = 0 for basic variables // But we explicitally skip setting basic variables here @@ -1931,9 +1997,11 @@ void set_primal_variables_on_bounds(const lp_problem_t& lp, } x[j] = lp.lower[j]; vstatus[j] = variable_status_t::NONBASIC_FIXED; - } else if (z[j] == 0 && lp.lower[j] > -inf && vstatus[j] == variable_status_t::NONBASIC_LOWER) { + } else if (z[j] >= -tol && lp.lower[j] > -inf && + vstatus[j] == variable_status_t::NONBASIC_LOWER) { x[j] = lp.lower[j]; - } else if (z[j] == 0 && lp.upper[j] < inf && vstatus[j] == variable_status_t::NONBASIC_UPPER) { + } else if (z[j] <= tol && lp.upper[j] < inf && + vstatus[j] == variable_status_t::NONBASIC_UPPER) { x[j] = lp.upper[j]; } else if (z[j] >= 0 && lp.lower[j] > -inf) { if (vstatus[j] != variable_status_t::NONBASIC_LOWER) { @@ -2018,7 +2086,9 @@ f_t amount_of_perturbation(const lp_problem_t& lp, const std::vector -void prepare_optimality(const lp_problem_t& lp, +void prepare_optimality(i_t info, + f_t orig_primal_infeas, + const lp_problem_t& lp, const simplex_solver_settings_t& settings, basis_update_mpf_t& ft, const std::vector& objective, @@ -2037,9 +2107,10 @@ void prepare_optimality(const lp_problem_t& lp, const i_t m = lp.num_rows; const i_t n = lp.num_cols; - sol.objective = compute_objective(lp, sol.x); - sol.user_objective = compute_user_objective(lp, sol.objective); - f_t perturbation = phase2::amount_of_perturbation(lp, objective); + sol.objective = compute_objective(lp, sol.x); + sol.user_objective = compute_user_objective(lp, sol.objective); + f_t perturbation = phase2::amount_of_perturbation(lp, objective); + f_t orig_perturbation = perturbation; if (perturbation > 1e-6 && phase == 2) { // Try to remove perturbation std::vector unperturbed_y(m); @@ -2079,6 +2150,27 @@ void prepare_optimality(const lp_problem_t& lp, settings.log.printf("Perturbation: %.2e\n", perturbation); } } + +#ifdef CHECK_PRIMAL_INFEASIBILITIES + if (primal_infeas > 10.0 * settings.primal_tol) { + f_t basic_infeas = 0.0; + f_t nonbasic_infeas = 0.0; + f_t basic_over = 0.0; + phase2::primal_infeasibility_breakdown( + lp, settings, vstatus, x, basic_infeas, nonbasic_infeas, basic_over); + settings.log.printf( + "Primal infeasibility %e/%e (Basic %e, Nonbasic %e, Basic over %e). Perturbation %e/%e. Info " + "%d\n", + primal_infeas, + orig_primal_infeas, + basic_infeas, + nonbasic_infeas, + basic_over, + orig_perturbation, + perturbation, + info); + } +#endif } template @@ -2319,12 +2411,19 @@ dual::status_t dual_phase2_with_advanced_basis(i_t phase, basic_list, nonbasic_list, delta_y_steepest_edge); } else { std::fill(delta_y_steepest_edge.begin(), delta_y_steepest_edge.end(), -1); - if (phase2::initialize_steepest_edge_norms( - lp, settings, start_time, basic_list, ft, delta_y_steepest_edge) == -1) { - return dual::status_t::TIME_LIMIT; - } + f_t steepest_edge_start = tic(); + i_t status = phase2::initialize_steepest_edge_norms( + lp, settings, start_time, basic_list, ft, delta_y_steepest_edge); + f_t steepest_edge_time = toc(steepest_edge_start); + if (status == CONCURRENT_HALT_RETURN) { return dual::status_t::CONCURRENT_LIMIT; } + if (status == -1) { return dual::status_t::TIME_LIMIT; } } } else { + // Check that none of the basic variables have a steepest edge that is nonpositive + for (i_t k = 0; k < m; k++) { + const i_t j = basic_list[k]; + if (delta_y_steepest_edge[j] <= 0.0) { delta_y_steepest_edge[j] = 1e-4; } + } settings.log.printf("using exisiting steepest edge %e\n", vector_norm2(delta_y_steepest_edge)); } @@ -2380,6 +2479,16 @@ dual::status_t dual_phase2_with_advanced_basis(i_t phase, i_t dense_delta_z = 0; phase2::phase2_timers_t timers(false); + if (phase == 2) { + settings.log.printf("%5d %+.16e %7d %.8e %.2e %.2f\n", + 0, + compute_user_objective(lp, obj), + infeasibility_indices.size(), + primal_infeasibility_squared, + 0.0, + toc(start_time)); + } + while (iter < iter_limit) { // Pricing i_t direction = 0; @@ -2405,7 +2514,67 @@ dual::status_t dual_phase2_with_advanced_basis(i_t phase, } timers.pricing_time += timers.stop_timer(); if (leaving_index == -1) { - phase2::prepare_optimality(lp, +#ifdef CHECK_BASIS_UPDATE + for (i_t k = 0; k < basic_list.size(); k++) { + const i_t jj = basic_list[k]; + sparse_vector_t ei_sparse(m, 1); + ei_sparse.i[0] = k; + ei_sparse.x[0] = 1.0; + sparse_vector_t ubar_sparse(m, 0); + ft.b_transpose_solve(ei_sparse, ubar_sparse); + std::vector ubar_dense(m); + ubar_sparse.to_dense(ubar_dense); + std::vector BTu_dense(m); + b_transpose_multiply(lp, basic_list, ubar_dense, BTu_dense); + for (i_t l = 0; l < m; l++) { + if (l != k) { + settings.log.printf("BTu_dense[%d] = %e i %d\n", l, BTu_dense[l], k); + } else { + settings.log.printf("BTu_dense[%d] = %e != 1.0 i %d\n", l, BTu_dense[l], k); + } + } + for (i_t h = 0; h < m; h++) { + settings.log.printf("i %d ubar_dense[%d] = %.16e\n", k, h, ubar_dense[h]); + } + } + settings.log.printf("ft.num_updates() %d\n", ft.num_updates()); + for (i_t h = 0; h < m; h++) { + settings.log.printf("basic_list[%d] = %d\n", h, basic_list[h]); + } + +#endif + +#ifdef CHECK_PRIMAL_INFEASIBILITIES + primal_infeasibility_squared = + phase2::compute_initial_primal_infeasibilities(lp, + settings, + basic_list, + x, + squared_infeasibilities, + infeasibility_indices, + primal_infeasibility); + if (primal_infeasibility > settings.primal_tol) { + const i_t nz = infeasibility_indices.size(); + for (i_t k = 0; k < nz; ++k) { + const i_t j = infeasibility_indices[k]; + const f_t squared_infeas = squared_infeasibilities[j]; + const f_t val = squared_infeas / delta_y_steepest_edge[j]; + if (squared_infeas >= 0.0 && delta_y_steepest_edge[j] < 0.0) { + settings.log.printf( + "Iter %d potential leaving %d val %e squared infeas %e delta_y_steepest_edge %e\n", + iter, + j, + val, + squared_infeas, + delta_y_steepest_edge[j]); + } + } + } +#endif + + phase2::prepare_optimality(0, + primal_infeasibility, + lp, settings, ft, objective, @@ -2527,7 +2696,7 @@ dual::status_t dual_phase2_with_advanced_basis(i_t phase, delta_z_indices, nonbasic_mark); entering_index = bfrt.compute_step_length(step_length, nonbasic_entering_index); - if (entering_index == -4) { + if (entering_index == RATIO_TEST_NUMERICAL_ISSUES) { settings.log.printf("Numerical issues encountered in ratio test.\n"); return dual::status_t::NUMERICAL; } @@ -2536,9 +2705,9 @@ dual::status_t dual_phase2_with_advanced_basis(i_t phase, entering_index = phase2::phase2_ratio_test( lp, settings, vstatus, nonbasic_list, z, delta_z, step_length, nonbasic_entering_index); } - if (entering_index == -2) { return dual::status_t::TIME_LIMIT; } - if (entering_index == -3) { return dual::status_t::CONCURRENT_LIMIT; } - if (entering_index == -1) { + if (entering_index == RATIO_TEST_TIME_LIMIT) { return dual::status_t::TIME_LIMIT; } + if (entering_index == CONCURRENT_HALT_RETURN) { return dual::status_t::CONCURRENT_LIMIT; } + if (entering_index == RATIO_TEST_NO_ENTERING_VARIABLE) { settings.log.printf("No entering variable found. Iter %d\n", iter); settings.log.printf("Scaled infeasibility %e\n", max_val); f_t perturbation = phase2::amount_of_perturbation(lp, objective); @@ -2577,7 +2746,9 @@ dual::status_t dual_phase2_with_advanced_basis(i_t phase, // Need to reset the objective value, since we have recomputed x obj = phase2::compute_perturbed_objective(objective, x); if (dual_infeas <= settings.dual_tol && primal_infeasibility <= settings.primal_tol) { - phase2::prepare_optimality(lp, + phase2::prepare_optimality(1, + primal_infeasibility, + lp, settings, ft, objective, @@ -2620,7 +2791,9 @@ dual::status_t dual_phase2_with_advanced_basis(i_t phase, if (primal_infeasibility <= settings.primal_tol && orig_dual_infeas <= settings.dual_tol) { - phase2::prepare_optimality(lp, + phase2::prepare_optimality(2, + primal_infeasibility, + lp, settings, ft, objective, @@ -2849,7 +3022,7 @@ dual::status_t dual_phase2_with_advanced_basis(i_t phase, squared_infeasibilities, infeasibility_indices, entering_index, - primal_infeasibility); + primal_infeasibility_squared); phase2::clean_up_infeasibilities(squared_infeasibilities, infeasibility_indices); diff --git a/cpp/src/dual_simplex/presolve.cpp b/cpp/src/dual_simplex/presolve.cpp index bbfe18d61d..9788968874 100644 --- a/cpp/src/dual_simplex/presolve.cpp +++ b/cpp/src/dual_simplex/presolve.cpp @@ -271,8 +271,6 @@ i_t convert_less_than_to_equal(const user_problem_t& user_problem, // We must convert rows in the form: a_i^T x <= beta // into: a_i^T x + s_i = beta, s_i >= 0 - csr_matrix_t Arow(0, 0, 0); - problem.A.to_compressed_row(Arow); i_t num_cols = problem.num_cols + less_rows; i_t nnz = problem.A.col_start[problem.num_cols] + less_rows; problem.A.col_start.resize(num_cols + 1); @@ -446,7 +444,7 @@ i_t find_dependent_rows(lp_problem_t& problem, std::vector q(m); i_t pivots = right_looking_lu_row_permutation_only(C, settings, 1e-13, tic(), q, pinv); - + if (pivots == CONCURRENT_HALT_RETURN) { return CONCURRENT_HALT_RETURN; } if (pivots < m) { settings.log.printf("Found %d dependent rows\n", m - pivots); const i_t num_dependent = m - pivots; @@ -629,8 +627,8 @@ void convert_user_problem(const user_problem_t& user_problem, // Empty var_types means that all variables are continuous bounds_strengthening_t strengthening(problem, Arow, row_sense, {}); - std::fill(strengthening.bounds_changed.begin(), strengthening.bounds_changed.end(), true); - strengthening.bounds_strengthening(problem.lower, problem.upper, settings); + std::vector bounds_changed(problem.num_cols, true); + strengthening.bounds_strengthening(settings, bounds_changed, problem.lower, problem.upper); } settings.log.debug( @@ -1101,6 +1099,7 @@ i_t presolve(const lp_problem_t& original, i_t infeasible; f_t dependent_row_start = tic(); const i_t independent_rows = find_dependent_rows(problem, settings, dependent_rows, infeasible); + if (independent_rows == CONCURRENT_HALT_RETURN) { return CONCURRENT_HALT_RETURN; } if (infeasible != kOk) { settings.log.printf("Found problem infeasible in presolve\n"); return -1; diff --git a/cpp/src/dual_simplex/primal.cpp b/cpp/src/dual_simplex/primal.cpp index 69f15ba188..38cddc0e24 100644 --- a/cpp/src/dual_simplex/primal.cpp +++ b/cpp/src/dual_simplex/primal.cpp @@ -296,7 +296,9 @@ primal::status_t primal_phase2(i_t phase, std::vector slacks_needed; i_t rank = factorize_basis(lp.A, settings, basic_list, L, U, p, pinv, q, deficient, slacks_needed); - if (rank != m) { + if (rank == CONCURRENT_HALT_RETURN) { + return primal::status_t::CONCURRENT_LIMIT; + } else if (rank != m) { settings.log.debug("Failed to factorize basis. rank %d m %d\n", rank, m); basis_repair(lp.A, settings, @@ -307,8 +309,10 @@ primal::status_t primal_phase2(i_t phase, basic_list, nonbasic_list, vstatus); - if (factorize_basis(lp.A, settings, basic_list, L, U, p, pinv, q, deficient, slacks_needed) == - -1) { + rank = factorize_basis(lp.A, settings, basic_list, L, U, p, pinv, q, deficient, slacks_needed); + if (rank == CONCURRENT_HALT_RETURN) { + return primal::status_t::CONCURRENT_LIMIT; + } else if (rank == -1) { settings.log.printf("Failed to factorize basis after repair. rank %d m %d\n", rank, m); return primal::status_t::NUMERICAL; } else { diff --git a/cpp/src/dual_simplex/primal.hpp b/cpp/src/dual_simplex/primal.hpp index a5d356fdbc..bd72fee4b3 100644 --- a/cpp/src/dual_simplex/primal.hpp +++ b/cpp/src/dual_simplex/primal.hpp @@ -1,6 +1,6 @@ /* clang-format off */ /* - * SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved. + * SPDX-FileCopyrightText: Copyright (c) 2025-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. * SPDX-License-Identifier: Apache-2.0 */ /* clang-format on */ @@ -23,7 +23,9 @@ enum class status_t { PRIMAL_UNBOUNDED = 1, NUMERICAL = 2, NOT_LOADED = 3, - ITERATION_LIMIT = 4 + TIME_LIMIT = 4, + ITERATION_LIMIT = 5, + CONCURRENT_LIMIT = 6 }; } diff --git a/cpp/src/dual_simplex/pseudo_costs.cpp b/cpp/src/dual_simplex/pseudo_costs.cpp index f56f3be4be..015034b775 100644 --- a/cpp/src/dual_simplex/pseudo_costs.cpp +++ b/cpp/src/dual_simplex/pseudo_costs.cpp @@ -210,9 +210,6 @@ static cuopt::mps_parser::mps_data_model_t simplex_problem_to_mps_data mps_model.set_constraint_lower_bounds(constraint_lower.data(), m); mps_model.set_constraint_upper_bounds(constraint_upper.data(), m); - - // TODO verify - // Set maximize flag (obj_scale: 1.0 for min, -1.0 for max) mps_model.set_maximize(user_problem.obj_scale < 0); return mps_model; @@ -236,14 +233,10 @@ void strong_branching(const user_problem_t& original_problem, pc.strong_branch_up.assign(fractional.size(), 0); pc.num_strong_branches_completed = 0; - settings.log.printf("Strong branching using %d threads and %ld fractional variables\n", - settings.num_threads, - fractional.size()); - if (settings.mip_batch_pdlp_strong_branching) { settings.log.printf("Batch PDLP strong branching enabled\n"); - std::chrono::steady_clock::time_point start_batch = std::chrono::steady_clock::now(); + f_t start_batch = tic(); // Use original_problem to create the BatchLP problem csr_matrix_t A_row(original_problem.A.m, original_problem.A.n, 0); @@ -263,8 +256,7 @@ void strong_branching(const user_problem_t& original_problem, const auto mps_model = simplex_problem_to_mps_data_model(original_problem); const auto solutions = batch_pdlp_solve(original_problem.handle_ptr, mps_model, fractional, fraction_values); - std::chrono::steady_clock::time_point end_batch = std::chrono::steady_clock::now(); - std::chrono::duration duration = end_batch - start_batch; + f_t batch_pdlp_strong_branching_time = toc(start_batch); // Find max iteration on how many are done accross the batch i_t max_iterations = 0; @@ -279,8 +271,8 @@ void strong_branching(const user_problem_t& original_problem, } settings.log.printf( - "Batch PDLP strong branching took %.2f seconds. Solved %d/%d with max %d iterations\n", - duration.count(), + "Batch PDLP strong branching completed in %.2fs. Solved %d/%d with max %d iterations\n", + batch_pdlp_strong_branching_time, amount_done, fractional.size() * 2, max_iterations); @@ -312,7 +304,10 @@ void strong_branching(const user_problem_t& original_problem, pc.strong_branch_up[k] = obj_up - root_obj; } } else { - std::chrono::steady_clock::time_point start_timea = std::chrono::steady_clock::now(); + settings.log.printf("Strong branching using %d threads and %ld fractional variables\n", + settings.num_threads, + fractional.size()); + f_t strong_branching_start_time = tic(); #pragma omp parallel num_threads(settings.num_threads) { @@ -349,9 +344,7 @@ void strong_branching(const user_problem_t& original_problem, pc); } } - std::chrono::steady_clock::time_point end_timea = std::chrono::steady_clock::now(); - std::chrono::duration duration = end_timea - start_timea; - settings.log.printf("Dual Simplex Strong branching took %.2f seconds\n", duration.count()); + settings.log.printf("Strong branching completed in %.2fs\n", toc(strong_branching_start_time)); } pc.update_pseudo_costs_from_strong_branching(fractional, root_soln); diff --git a/cpp/src/dual_simplex/right_looking_lu.cpp b/cpp/src/dual_simplex/right_looking_lu.cpp index a63c1181fa..82ea7c0e10 100644 --- a/cpp/src/dual_simplex/right_looking_lu.cpp +++ b/cpp/src/dual_simplex/right_looking_lu.cpp @@ -1,6 +1,6 @@ /* clang-format off */ /* - * SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved. + * SPDX-FileCopyrightText: Copyright (c) 2025-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. * SPDX-License-Identifier: Apache-2.0 */ /* clang-format on */ @@ -1115,7 +1115,7 @@ i_t right_looking_lu_row_permutation_only(const csc_matrix_t& A, if (settings.concurrent_halt != nullptr && *settings.concurrent_halt == 1) { settings.log.printf("Concurrent halt\n"); - return -2; + return CONCURRENT_HALT_RETURN; } } diff --git a/cpp/src/dual_simplex/simplex_solver_settings.hpp b/cpp/src/dual_simplex/simplex_solver_settings.hpp index cbe0cbba5e..f9911ee53a 100644 --- a/cpp/src/dual_simplex/simplex_solver_settings.hpp +++ b/cpp/src/dual_simplex/simplex_solver_settings.hpp @@ -87,9 +87,19 @@ struct simplex_solver_settings_t { iteration_log_frequency(1000), first_iteration_log(2), num_threads(omp_get_max_threads() - 1), + max_cut_passes(0), + mir_cuts(-1), + mixed_integer_gomory_cuts(-1), + knapsack_cuts(-1), + strong_chvatal_gomory_cuts(-1), + reduced_cost_strengthening(-1), + cut_change_threshold(1e-3), + cut_min_orthogonality(0.5), num_bfs_workers(std::max(num_threads / 4, 1)), random_seed(0), inside_mip(0), + sub_mip(0), + reliability_branching(-1), solution_callback(nullptr), heuristic_preemption_callback(nullptr), concurrent_halt(nullptr) @@ -154,6 +164,17 @@ struct simplex_solver_settings_t { i_t first_iteration_log; // number of iterations to log at beginning of solve i_t num_threads; // number of threads to use i_t random_seed; // random seed + i_t max_cut_passes; // number of cut passes to make + i_t mir_cuts; // -1 automatic, 0 to disable, >0 to enable MIR cuts + i_t mixed_integer_gomory_cuts; // -1 automatic, 0 to disable, >0 to enable mixed integer Gomory + // cuts + i_t knapsack_cuts; // -1 automatic, 0 to disable, >0 to enable knapsack cuts + i_t strong_chvatal_gomory_cuts; // -1 automatic, 0 to disable, >0 to enable strong Chvatal Gomory + // cuts + i_t reduced_cost_strengthening; // -1 automatic, 0 to disable, >0 to enable reduced cost + // strengthening + f_t cut_change_threshold; // threshold for cut change + f_t cut_min_orthogonality; // minimum orthogonality for cuts i_t num_bfs_workers; // number of threads dedicated to the best-first search i_t mip_batch_pdlp_strong_branching{0}; // 0 if not using batch PDLP for strong branching, 1 if // using batch PDLP for strong branching @@ -161,6 +182,8 @@ struct simplex_solver_settings_t { diving_heuristics_settings_t diving_settings; // Settings for the diving heuristics i_t inside_mip; // 0 if outside MIP, 1 if inside MIP at root node, 2 if inside MIP at leaf node + i_t sub_mip; // 0 if in regular MIP solve, 1 if in sub-MIP solve + i_t reliability_branching; // -1 automatic, 0 to disable, >0 to enable reliability branching std::function&, f_t)> solution_callback; std::function&, f_t)> node_processed_callback; std::function heuristic_preemption_callback; diff --git a/cpp/src/dual_simplex/solution.hpp b/cpp/src/dual_simplex/solution.hpp index d1d745cbd9..a678e2fd70 100644 --- a/cpp/src/dual_simplex/solution.hpp +++ b/cpp/src/dual_simplex/solution.hpp @@ -1,6 +1,6 @@ /* clang-format off */ /* - * SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved. + * SPDX-FileCopyrightText: Copyright (c) 2025-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. * SPDX-License-Identifier: Apache-2.0 */ /* clang-format on */ @@ -39,7 +39,7 @@ class lp_solution_t { std::vector x; // Dual solution vector. Lagrange multipliers for equality constraints. std::vector y; - // Dual solution vector. Lagrange multipliers for inequality constraints. + // Reduced costs std::vector z; f_t objective; f_t user_objective; diff --git a/cpp/src/dual_simplex/solve.cpp b/cpp/src/dual_simplex/solve.cpp index 1f31a757d4..37297d9beb 100644 --- a/cpp/src/dual_simplex/solve.cpp +++ b/cpp/src/dual_simplex/solve.cpp @@ -1,6 +1,6 @@ /* clang-format off */ /* - * SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved. + * SPDX-FileCopyrightText: Copyright (c) 2025-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. * SPDX-License-Identifier: Apache-2.0 */ /* clang-format on */ @@ -8,6 +8,7 @@ #include #include +#include #include #include #include @@ -140,6 +141,7 @@ lp_status_t solve_linear_program_with_advanced_basis( lp_problem_t presolved_lp(original_lp.handle_ptr, 1, 1, 1); presolve_info_t presolve_info; const i_t ok = presolve(original_lp, settings, presolved_lp, presolve_info); + if (ok == CONCURRENT_HALT_RETURN) { return lp_status_t::CONCURRENT_LIMIT; } if (ok == -1) { return lp_status_t::INFEASIBLE; } constexpr bool write_out_matlab = false; @@ -314,6 +316,7 @@ lp_status_t solve_linear_program_with_barrier(const user_problem_t& us presolve_info_t presolve_info; lp_problem_t presolved_lp(user_problem.handle_ptr, 1, 1, 1); const i_t ok = presolve(original_lp, barrier_settings, presolved_lp, presolve_info); + if (ok == CONCURRENT_HALT_RETURN) { return lp_status_t::CONCURRENT_LIMIT; } if (ok == -1) { return lp_status_t::INFEASIBLE; } // Apply columns scaling to the presolve LP @@ -457,18 +460,59 @@ lp_status_t solve_linear_program_with_barrier(const user_problem_t& us settings.log.printf("Primal objective: %e\n", dot(dualize_info.primal_problem.objective, primal_solution.x)); - std::vector primal_residual = dualize_info.primal_problem.rhs; - matrix_vector_multiply( - dualize_info.primal_problem.A, 1.0, primal_solution.x, -1.0, primal_residual); std::vector inequality_rows(dualize_info.primal_problem.num_rows, 1); for (i_t i : dualize_info.equality_rows) { inequality_rows[i] = 0; } + i_t less_rows = 0; for (i_t i = 0; i < dualize_info.primal_problem.num_rows; ++i) { - if (inequality_rows[i] == 1) { - primal_residual[i] = std::max(primal_residual[i], 0.0); // a_i^T x - b_i <= 0 + if (inequality_rows[i] == 1) { less_rows++; } + } + // Add slack variables to the primal problem + if (less_rows > 0) { + std::vector slack_info = dualize_info.primal_problem.rhs; + matrix_vector_multiply( + dualize_info.primal_problem.A, -1.0, primal_solution.x, 1.0, slack_info); + + lp_problem_t& problem = dualize_info.primal_problem; + i_t num_cols = problem.num_cols + less_rows; + i_t nnz = problem.A.col_start[problem.num_cols] + less_rows; + problem.A.col_start.resize(num_cols + 1); + problem.A.i.resize(nnz); + problem.A.x.resize(nnz); + problem.lower.resize(num_cols); + problem.upper.resize(num_cols); + problem.objective.resize(num_cols); + primal_solution.x.resize(num_cols); + primal_solution.z.resize(num_cols); + + i_t p = problem.A.col_start[problem.num_cols]; + i_t j = problem.num_cols; + for (i_t i = 0; i < problem.num_rows; i++) { + if (inequality_rows[i] == 1) { + problem.lower[j] = 0.0; + problem.upper[j] = INFINITY; + problem.objective[j] = 0.0; + problem.A.i[p] = i; + problem.A.x[p] = 1.0; + primal_solution.x[j] = slack_info[i]; + primal_solution.z[j] = -primal_solution.y[i]; + problem.A.col_start[j++] = p++; + inequality_rows[i] = 0; + less_rows--; + } } + problem.A.col_start[num_cols] = p; + assert(less_rows == 0); + assert(p == nnz); + problem.A.n = num_cols; + problem.num_cols = num_cols; } + + std::vector primal_residual = dualize_info.primal_problem.rhs; + matrix_vector_multiply( + dualize_info.primal_problem.A, 1.0, primal_solution.x, -1.0, primal_residual); + f_t primal_residual_norm = vector_norm_inf(primal_residual); const f_t norm_b = vector_norm_inf(dualize_info.primal_problem.rhs); f_t primal_relative_residual = primal_residual_norm / (1.0 + norm_b); @@ -505,6 +549,13 @@ lp_status_t solve_linear_program_with_barrier(const user_problem_t& us if (!settings.crossover || barrier_lp.Q.n > 0) { return barrier_status; } if (settings.crossover && barrier_status == lp_status_t::OPTIMAL) { + { + std::vector rhs = original_lp.rhs; + matrix_vector_multiply(original_lp.A, 1.0, lp_solution.x, -1.0, rhs); + f_t primal_residual = vector_norm_inf(rhs); + settings.log.printf("Primal residual before adding artificial variables: %e\n", + primal_residual); + } // Check to see if we need to add artifical variables std::vector artificial_variables; artificial_variables.reserve(original_lp.num_rows); @@ -550,6 +601,12 @@ lp_status_t solve_linear_program_with_barrier(const user_problem_t& us lp_solution.x.size(), lp_solution.z.size()); #endif + + std::vector rhs = original_lp.rhs; + matrix_vector_multiply(original_lp.A, 1.0, lp_solution.x, -1.0, rhs); + f_t primal_residual = vector_norm_inf(rhs); + settings.log.printf("Primal residual after adding artificial variables: %e\n", + primal_residual); } // Run crossover @@ -597,7 +654,7 @@ i_t solve(const user_problem_t& problem, { i_t status; if (is_mip(problem) && !settings.relaxation) { - branch_and_bound_t branch_and_bound(problem, settings); + branch_and_bound_t branch_and_bound(problem, settings, tic()); mip_solution_t mip_solution(problem.num_cols); mip_status_t mip_status = branch_and_bound.solve(mip_solution); if (mip_status == mip_status_t::OPTIMAL) { @@ -636,7 +693,7 @@ i_t solve_mip_with_guess(const user_problem_t& problem, { i_t status; if (is_mip(problem)) { - branch_and_bound_t branch_and_bound(problem, settings); + branch_and_bound_t branch_and_bound(problem, settings, tic()); branch_and_bound.set_initial_guess(guess); mip_status_t mip_status = branch_and_bound.solve(solution); if (mip_status == mip_status_t::OPTIMAL) { diff --git a/cpp/src/dual_simplex/sparse_cholesky.cuh b/cpp/src/dual_simplex/sparse_cholesky.cuh index eea7ee8f32..e2b351afaa 100644 --- a/cpp/src/dual_simplex/sparse_cholesky.cuh +++ b/cpp/src/dual_simplex/sparse_cholesky.cuh @@ -67,7 +67,7 @@ class sparse_cholesky_base_t { "details: " #msg "\n", \ status); \ CUDSS_EXAMPLE_FREE; \ - return -2; \ + return -1; \ } \ } while (0); @@ -444,7 +444,9 @@ class sparse_cholesky_cudss_t : public sparse_cholesky_base_t { raft::common::nvtx::range fun_scope("Barrier: cuDSS Analyze : CUDSS_PHASE_ANALYSIS"); status = cudssExecute(handle, CUDSS_PHASE_REORDERING, solverConfig, solverData, A, cudss_x, cudss_b); - if (settings_.concurrent_halt != nullptr && *settings_.concurrent_halt == 1) { return -2; } + if (settings_.concurrent_halt != nullptr && *settings_.concurrent_halt == 1) { + return CONCURRENT_HALT_RETURN; + } if (status != CUDSS_STATUS_SUCCESS) { settings_.log.printf( "FAILED: CUDSS call ended unsuccessfully with status = %d, details: cuDSSExecute for " @@ -458,7 +460,9 @@ class sparse_cholesky_cudss_t : public sparse_cholesky_base_t { status = cudssExecute( handle, CUDSS_PHASE_SYMBOLIC_FACTORIZATION, solverConfig, solverData, A, cudss_x, cudss_b); - if (settings_.concurrent_halt != nullptr && *settings_.concurrent_halt == 1) { return -2; } + if (settings_.concurrent_halt != nullptr && *settings_.concurrent_halt == 1) { + return CONCURRENT_HALT_RETURN; + } if (status != CUDSS_STATUS_SUCCESS) { settings_.log.printf( "FAILED: CUDSS call ended unsuccessfully with status = %d, details: cuDSSExecute for " @@ -514,7 +518,9 @@ class sparse_cholesky_cudss_t : public sparse_cholesky_base_t { f_t start_numeric = tic(); status = cudssExecute( handle, CUDSS_PHASE_FACTORIZATION, solverConfig, solverData, A, cudss_x, cudss_b); - if (settings_.concurrent_halt != nullptr && *settings_.concurrent_halt == 1) { return -2; } + if (settings_.concurrent_halt != nullptr && *settings_.concurrent_halt == 1) { + return CONCURRENT_HALT_RETURN; + } if (status != CUDSS_STATUS_SUCCESS) { settings_.log.printf( "FAILED: CUDSS call ended unsuccessfully with status = %d, details: cuDSSExecute for " @@ -528,7 +534,9 @@ class sparse_cholesky_cudss_t : public sparse_cholesky_base_t { #endif f_t numeric_time = toc(start_numeric); - if (settings_.concurrent_halt != nullptr && *settings_.concurrent_halt == 1) { return -2; } + if (settings_.concurrent_halt != nullptr && *settings_.concurrent_halt == 1) { + return CONCURRENT_HALT_RETURN; + } int info; size_t sizeWritten = 0; @@ -626,7 +634,9 @@ class sparse_cholesky_cudss_t : public sparse_cholesky_base_t { A_created = true; // Perform symbolic analysis - if (settings_.concurrent_halt != nullptr && *settings_.concurrent_halt == 1) { return -2; } + if (settings_.concurrent_halt != nullptr && *settings_.concurrent_halt == 1) { + return CONCURRENT_HALT_RETURN; + } f_t start_analysis = tic(); CUDSS_CALL_AND_CHECK( cudssExecute(handle, CUDSS_PHASE_REORDERING, solverConfig, solverData, A, cudss_x, cudss_b), @@ -634,7 +644,9 @@ class sparse_cholesky_cudss_t : public sparse_cholesky_base_t { "cudssExecute for reordering"); f_t reorder_time = toc(start_analysis); - if (settings_.concurrent_halt != nullptr && *settings_.concurrent_halt == 1) { return -2; } + if (settings_.concurrent_halt != nullptr && *settings_.concurrent_halt == 1) { + return CONCURRENT_HALT_RETURN; + } f_t start_symbolic = tic(); @@ -650,7 +662,7 @@ class sparse_cholesky_cudss_t : public sparse_cholesky_base_t { if (settings_.concurrent_halt != nullptr && *settings_.concurrent_halt == 1) { RAFT_CUDA_TRY(cudaStreamSynchronize(stream)); handle_ptr_->get_stream().synchronize(); - return -2; + return CONCURRENT_HALT_RETURN; } int64_t lu_nz = 0; size_t size_written = 0; @@ -698,7 +710,9 @@ class sparse_cholesky_cudss_t : public sparse_cholesky_base_t { "cudssExecute for factorization"); f_t numeric_time = toc(start_numeric); - if (settings_.concurrent_halt != nullptr && *settings_.concurrent_halt == 1) { return -2; } + if (settings_.concurrent_halt != nullptr && *settings_.concurrent_halt == 1) { + return CONCURRENT_HALT_RETURN; + } int info; size_t sizeWritten = 0; @@ -761,7 +775,9 @@ class sparse_cholesky_cudss_t : public sparse_cholesky_base_t { cudssMatrixSetValues(cudss_x, x.data()), status, "cudssMatrixSetValues for x"); status = cudssExecute(handle, CUDSS_PHASE_SOLVE, solverConfig, solverData, A, cudss_x, cudss_b); - if (settings_.concurrent_halt != nullptr && *settings_.concurrent_halt == 1) { return -2; } + if (settings_.concurrent_halt != nullptr && *settings_.concurrent_halt == 1) { + return CONCURRENT_HALT_RETURN; + } if (status != CUDSS_STATUS_SUCCESS) { settings_.log.printf( "FAILED: CUDSS call ended unsuccessfully with status = %d, details: cuDSSExecute for " diff --git a/cpp/src/dual_simplex/sparse_matrix.cpp b/cpp/src/dual_simplex/sparse_matrix.cpp index 7edc7b1eb5..1d8f12a3a5 100644 --- a/cpp/src/dual_simplex/sparse_matrix.cpp +++ b/cpp/src/dual_simplex/sparse_matrix.cpp @@ -1,6 +1,6 @@ /* clang-format off */ /* - * SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved. + * SPDX-FileCopyrightText: Copyright (c) 2025-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. * SPDX-License-Identifier: Apache-2.0 */ /* clang-format on */ @@ -363,6 +363,75 @@ i_t csc_matrix_t::remove_row(i_t row) return 0; } +template +i_t csr_matrix_t::append_rows(const csr_matrix_t& C) +{ + const i_t old_m = this->m; + const i_t n = this->n; + const i_t old_nz = this->row_start[old_m]; + const i_t C_row = C.m; + if (C.n > n) { + printf("append_rows error: C.n %d n %d\n", C.n, n); + return -1; + } + const i_t C_nz = C.row_start[C_row]; + const i_t new_nz = old_nz + C_nz; + const i_t new_m = old_m + C_row; + + this->j.resize(new_nz); + this->x.resize(new_nz); + this->row_start.resize(new_m + 1); + + i_t nz = old_nz; + for (i_t i = old_m; i < new_m; i++) { + const i_t k = i - old_m; + const i_t nz_row = C.row_start[k + 1] - C.row_start[k]; + this->row_start[i] = nz; + nz += nz_row; + } + this->row_start[new_m] = nz; + + for (i_t p = old_nz; p < new_nz; p++) { + const i_t q = p - old_nz; + this->j[p] = C.j[q]; + } + + for (i_t p = old_nz; p < new_nz; p++) { + const i_t q = p - old_nz; + this->x[p] = C.x[q]; + } + + this->m = new_m; + this->nz_max = new_nz; + return 0; +} + +template +i_t csr_matrix_t::append_row(const sparse_vector_t& c) +{ + const i_t old_m = this->m; + const i_t old_nz = this->row_start[old_m]; + const i_t c_nz = c.i.size(); + const i_t new_nz = old_nz + c_nz; + const i_t new_m = old_m + 1; + + this->j.resize(new_nz); + this->x.resize(new_nz); + this->row_start.resize(new_m + 1); + this->row_start[new_m] = new_nz; + + i_t nz = old_nz; + for (i_t k = 0; k < c_nz; k++) { + this->j[nz] = c.i[k]; + this->x[nz] = c.x[k]; + nz++; + } + + this->m = new_m; + this->nz_max = new_nz; + return 0; +} + template void csc_matrix_t::print_matrix(FILE* fid) const { @@ -505,6 +574,10 @@ i_t csc_matrix_t::check_matrix(std::string matrix_name) const #ifdef CHECK_MATRIX std::vector row_marker(this->m, -1); for (i_t j = 0; j < this->n; ++j) { + if (j >= col_start.size()) { + printf("Col start too small size %ld n %d\n", col_start.size(), this->n); + return -1; + } const i_t col_start = this->col_start[j]; const i_t col_end = this->col_start[j + 1]; if (col_start > col_end || col_start > this->col_start[this->n]) { @@ -559,7 +632,7 @@ size_t csc_matrix_t::hash() const } template -void csr_matrix_t::check_matrix(std::string matrix_name) const +i_t csr_matrix_t::check_matrix(std::string matrix_name) const { std::vector col_marker(this->n, -1); for (i_t i = 0; i < this->m; ++i) { @@ -567,12 +640,18 @@ void csr_matrix_t::check_matrix(std::string matrix_name) const const i_t row_end = this->row_start[i + 1]; for (i_t p = row_start; p < row_end; ++p) { const i_t j = this->j[p]; + if (j < 0 || j >= this->n) { + printf("CSR Error: column index %d not in range [0, %d)\n", j, this->n); + return -1; + } if (col_marker[j] == i) { printf("CSR Error (%s) : repeated column index %d in row %d\n", matrix_name.c_str(), j, i); + return -1; } col_marker[j] = i; } } + return 0; } // x <- x + alpha * A(:, j) diff --git a/cpp/src/dual_simplex/sparse_matrix.hpp b/cpp/src/dual_simplex/sparse_matrix.hpp index 9ae8ea80be..0b6c0b11d6 100644 --- a/cpp/src/dual_simplex/sparse_matrix.hpp +++ b/cpp/src/dual_simplex/sparse_matrix.hpp @@ -1,6 +1,6 @@ /* clang-format off */ /* - * SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved. + * SPDX-FileCopyrightText: Copyright (c) 2025-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. * SPDX-License-Identifier: Apache-2.0 */ /* clang-format on */ @@ -151,8 +151,14 @@ class csr_matrix_t { // Create a new matrix with the marked rows removed i_t remove_rows(std::vector& row_marker, csr_matrix_t& Aout) const; + // Append rows from another CSR matrix + i_t append_rows(const csr_matrix_t& C); + + // Append a row from a sparse vector + i_t append_row(const sparse_vector_t& c); + // Ensures no repeated column indices within a row - void check_matrix(std::string matrix_name = "") const; + i_t check_matrix(std::string matrix_name = "") const; bool is_diagonal() const { diff --git a/cpp/src/dual_simplex/sparse_vector.cpp b/cpp/src/dual_simplex/sparse_vector.cpp index 2d47456505..4e2ecfa196 100644 --- a/cpp/src/dual_simplex/sparse_vector.cpp +++ b/cpp/src/dual_simplex/sparse_vector.cpp @@ -1,6 +1,6 @@ /* clang-format off */ /* - * SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved. + * SPDX-FileCopyrightText: Copyright (c) 2025-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. * SPDX-License-Identifier: Apache-2.0 */ /* clang-format on */ @@ -28,6 +28,21 @@ sparse_vector_t::sparse_vector_t(const csc_matrix_t& A, i_t } } +template +sparse_vector_t::sparse_vector_t(const csr_matrix_t& A, i_t row) +{ + const i_t row_start = A.row_start[row]; + const i_t row_end = A.row_start[row + 1]; + const i_t nz = row_end - row_start; + n = A.n; + i.reserve(nz); + x.reserve(nz); + for (i_t k = row_start; k < row_end; ++k) { + i.push_back(A.j[k]); + x.push_back(A.x[k]); + } +} + template void sparse_vector_t::from_dense(const std::vector& in) { @@ -106,6 +121,17 @@ void sparse_vector_t::inverse_permute_vector(const std::vector& p y.i = i_perm; } +template +f_t sparse_vector_t::dot(const std::vector& x_dense) const +{ + const i_t nz = i.size(); + f_t dot = 0.0; + for (i_t k = 0; k < nz; ++k) { + dot += x[k] * x_dense[i[k]]; + } + return dot; +} + template f_t sparse_vector_t::sparse_dot(const csc_matrix_t& Y, i_t y_col) const { @@ -207,6 +233,28 @@ f_t sparse_vector_t::find_coefficient(i_t index) const return std::numeric_limits::quiet_NaN(); } +template +void sparse_vector_t::squeeze(sparse_vector_t& y) const +{ + y.n = n; + + i_t nz = 0; + const i_t nx = x.size(); + for (i_t k = 0; k < nx; k++) { + if (x[k] != 0.0) { nz++; } + } + y.i.reserve(nz); + y.x.reserve(nz); + y.i.clear(); + y.x.clear(); + for (i_t k = 0; k < nx; k++) { + if (x[k] != 0.0) { + y.i.push_back(i[k]); + y.x.push_back(x[k]); + } + } +} + #ifdef DUAL_SIMPLEX_INSTANTIATE_DOUBLE template class sparse_vector_t; #endif diff --git a/cpp/src/dual_simplex/sparse_vector.hpp b/cpp/src/dual_simplex/sparse_vector.hpp index 7acfdc8b5e..95e9afa29e 100644 --- a/cpp/src/dual_simplex/sparse_vector.hpp +++ b/cpp/src/dual_simplex/sparse_vector.hpp @@ -1,6 +1,6 @@ /* clang-format off */ /* - * SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved. + * SPDX-FileCopyrightText: Copyright (c) 2025-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. * SPDX-License-Identifier: Apache-2.0 */ /* clang-format on */ @@ -25,6 +25,8 @@ class sparse_vector_t { sparse_vector_t(const std::vector& in) { from_dense(in); } // Construct a sparse vector from a column of a CSC matrix sparse_vector_t(const csc_matrix_t& A, i_t col); + // Construct a sparse vector from a row of a CSR matrix + sparse_vector_t(const csr_matrix_t& A, i_t row); // gather a dense vector into a sparse vector void from_dense(const std::vector& in); // convert a sparse vector into a CSC matrix with a single column @@ -38,6 +40,8 @@ class sparse_vector_t { void inverse_permute_vector(const std::vector& p); // inverse permute a sparse vector into another sparse vector void inverse_permute_vector(const std::vector& p, sparse_vector_t& y) const; + // compute the dot product of a sparse vector with a dense vector + f_t dot(const std::vector& x) const; // compute the dot product of a sparse vector with a column of a CSC matrix f_t sparse_dot(const csc_matrix_t& Y, i_t y_col) const; // ensure the coefficients in the sparse vectory are sorted in terms of increasing index @@ -47,6 +51,8 @@ class sparse_vector_t { void negate(); f_t find_coefficient(i_t index) const; + void squeeze(sparse_vector_t& y) const; + i_t n; std::vector i; std::vector x; diff --git a/cpp/src/dual_simplex/types.hpp b/cpp/src/dual_simplex/types.hpp index a3e15190e2..9de33ed3b3 100644 --- a/cpp/src/dual_simplex/types.hpp +++ b/cpp/src/dual_simplex/types.hpp @@ -1,6 +1,6 @@ /* clang-format off */ /* - * SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved. + * SPDX-FileCopyrightText: Copyright (c) 2025-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. * SPDX-License-Identifier: Apache-2.0 */ /* clang-format on */ @@ -19,4 +19,7 @@ using float64_t = double; constexpr float64_t inf = std::numeric_limits::infinity(); +// We return this constant to signal that a concurrent halt has occurred +#define CONCURRENT_HALT_RETURN -2 + } // namespace cuopt::linear_programming::dual_simplex diff --git a/cpp/src/math_optimization/solver_settings.cu b/cpp/src/math_optimization/solver_settings.cu index 493e730fb9..9dc6ac9c5e 100644 --- a/cpp/src/math_optimization/solver_settings.cu +++ b/cpp/src/math_optimization/solver_settings.cu @@ -72,7 +72,9 @@ solver_settings_t::solver_settings_t() : pdlp_settings(), mip_settings {CUOPT_MIP_ABSOLUTE_GAP, &mip_settings.tolerances.absolute_mip_gap, 0.0, CUOPT_INFINITY, 1e-10}, {CUOPT_MIP_RELATIVE_GAP, &mip_settings.tolerances.relative_mip_gap, 0.0, 1e-1, 1e-4}, {CUOPT_PRIMAL_INFEASIBLE_TOLERANCE, &pdlp_settings.tolerances.primal_infeasible_tolerance, 0.0, 1e-1, 1e-10}, - {CUOPT_DUAL_INFEASIBLE_TOLERANCE, &pdlp_settings.tolerances.dual_infeasible_tolerance, 0.0, 1e-1, 1e-10} + {CUOPT_DUAL_INFEASIBLE_TOLERANCE, &pdlp_settings.tolerances.dual_infeasible_tolerance, 0.0, 1e-1, 1e-10}, + {CUOPT_MIP_CUT_CHANGE_THRESHOLD, &mip_settings.cut_change_threshold, 0.0, std::numeric_limits::infinity(), 1e-3}, + {CUOPT_MIP_CUT_MIN_ORTHOGONALITY, &mip_settings.cut_min_orthogonality, 0.0, 1.0, 0.5} }; // Int parameters @@ -87,6 +89,12 @@ solver_settings_t::solver_settings_t() : pdlp_settings(), mip_settings {CUOPT_DUALIZE, &pdlp_settings.dualize, -1, 1, -1}, {CUOPT_ORDERING, &pdlp_settings.ordering, -1, 1, -1}, {CUOPT_BARRIER_DUAL_INITIAL_POINT, &pdlp_settings.barrier_dual_initial_point, -1, 1, -1}, + {CUOPT_MIP_CUT_PASSES, &mip_settings.max_cut_passes, -1, std::numeric_limits::max(), 10}, + {CUOPT_MIP_MIXED_INTEGER_ROUNDING_CUTS, &mip_settings.mir_cuts, -1, 1, -1}, + {CUOPT_MIP_MIXED_INTEGER_GOMORY_CUTS, &mip_settings.mixed_integer_gomory_cuts, -1, 1, -1}, + {CUOPT_MIP_KNAPSACK_CUTS, &mip_settings.knapsack_cuts, -1, 1, -1}, + {CUOPT_MIP_STRONG_CHVATAL_GOMORY_CUTS, &mip_settings.strong_chvatal_gomory_cuts, -1, 1, -1}, + {CUOPT_MIP_REDUCED_COST_STRENGTHENING, &mip_settings.reduced_cost_strengthening, -1, std::numeric_limits::max(), -1}, {CUOPT_NUM_GPUS, &pdlp_settings.num_gpus, 1, 2, 1}, {CUOPT_NUM_GPUS, &mip_settings.num_gpus, 1, 2, 1}, {CUOPT_MIP_BATCH_PDLP_STRONG_BRANCHING, &mip_settings.mip_batch_pdlp_strong_branching, 0, 1, 0}, diff --git a/cpp/src/mip/diversity/diversity_manager.cu b/cpp/src/mip/diversity/diversity_manager.cu index cf2180801d..f016753270 100644 --- a/cpp/src/mip/diversity/diversity_manager.cu +++ b/cpp/src/mip/diversity/diversity_manager.cu @@ -216,10 +216,11 @@ bool diversity_manager_t::run_presolve(f_t time_limit) lp_dual_optimal_solution.resize(problem_ptr->n_constraints, problem_ptr->handle_ptr->get_stream()); problem_ptr->handle_ptr->sync_stream(); - CUOPT_LOG_INFO("After trivial presolve: %d constraints, %d variables, objective offset %f.", + CUOPT_LOG_INFO("After cuOpt presolve: %d constraints, %d variables, objective offset %f.", problem_ptr->n_constraints, problem_ptr->n_variables, problem_ptr->presolve_data.objective_offset); + CUOPT_LOG_INFO("cuOpt presolve time: %.2f", stats.presolve_time); return true; } diff --git a/cpp/src/mip/diversity/lns/rins.cu b/cpp/src/mip/diversity/lns/rins.cu index af992d2e5c..cb086e1456 100644 --- a/cpp/src/mip/diversity/lns/rins.cu +++ b/cpp/src/mip/diversity/lns/rins.cu @@ -22,6 +22,8 @@ #include #include +#include + namespace cuopt::linear_programming::detail { template rins_t::rins_t(mip_solver_context_t& context_, @@ -260,6 +262,8 @@ void rins_t::run_rins() branch_and_bound_settings.integer_tol = context.settings.tolerances.integrality_tolerance; branch_and_bound_settings.num_threads = 2; branch_and_bound_settings.num_bfs_workers = 1; + branch_and_bound_settings.max_cut_passes = 0; + branch_and_bound_settings.sub_mip = 1; // In the future, let RINS use all the diving heuristics. For now, // restricting to guided diving. @@ -273,8 +277,8 @@ void rins_t::run_rins() f_t objective) { rins_solution_queue.push_back(solution); }; - dual_simplex::branch_and_bound_t branch_and_bound(branch_and_bound_problem, - branch_and_bound_settings); + dual_simplex::branch_and_bound_t branch_and_bound( + branch_and_bound_problem, branch_and_bound_settings, dual_simplex::tic()); branch_and_bound.set_initial_guess(cuopt::host_copy(fixed_assignment, rins_handle.get_stream())); branch_and_bound_status = branch_and_bound.solve(branch_and_bound_solution); diff --git a/cpp/src/mip/diversity/recombiners/sub_mip.cuh b/cpp/src/mip/diversity/recombiners/sub_mip.cuh index 00cd0730da..e636e74714 100644 --- a/cpp/src/mip/diversity/recombiners/sub_mip.cuh +++ b/cpp/src/mip/diversity/recombiners/sub_mip.cuh @@ -13,6 +13,7 @@ #include #include #include +#include namespace cuopt::linear_programming::detail { @@ -106,6 +107,8 @@ class sub_mip_recombiner_t : public recombiner_t { branch_and_bound_settings.integer_tol = context.settings.tolerances.integrality_tolerance; branch_and_bound_settings.num_threads = 2; branch_and_bound_settings.num_bfs_workers = 1; + branch_and_bound_settings.max_cut_passes = 0; + branch_and_bound_settings.sub_mip = 1; // In the future, let SubMIP use all the diving heuristics. For now, // restricting to guided diving. @@ -120,8 +123,8 @@ class sub_mip_recombiner_t : public recombiner_t { // disable B&B logs, so that it is not interfering with the main B&B thread branch_and_bound_settings.log.log = false; - dual_simplex::branch_and_bound_t branch_and_bound(branch_and_bound_problem, - branch_and_bound_settings); + dual_simplex::branch_and_bound_t branch_and_bound( + branch_and_bound_problem, branch_and_bound_settings, dual_simplex::tic()); branch_and_bound_status = branch_and_bound.solve(branch_and_bound_solution); if (solution_vector.size() > 0) { cuopt_assert(fixed_assignment.size() == branch_and_bound_solution.x.size(), diff --git a/cpp/src/mip/presolve/third_party_presolve.cpp b/cpp/src/mip/presolve/third_party_presolve.cpp index 9a212ebab0..b60747c57d 100644 --- a/cpp/src/mip/presolve/third_party_presolve.cpp +++ b/cpp/src/mip/presolve/third_party_presolve.cpp @@ -303,7 +303,7 @@ void check_presolve_status(const papilo::PresolveStatus& status) void check_postsolve_status(const papilo::PostsolveStatus& status) { switch (status) { - case papilo::PostsolveStatus::kOk: CUOPT_LOG_INFO("Post-solve status: succeeded"); break; + case papilo::PostsolveStatus::kOk: CUOPT_LOG_DEBUG("Post-solve status: succeeded"); break; case papilo::PostsolveStatus::kFailed: CUOPT_LOG_INFO( "Post-solve status: Post solved solution violates constraints. This is most likely due to " @@ -409,7 +409,7 @@ std::optional> third_party_presolve_t presolver; set_presolve_methods(presolver, category, dual_postsolve); diff --git a/cpp/src/mip/solve.cu b/cpp/src/mip/solve.cu index 62ee0bb95e..ee852fb291 100644 --- a/cpp/src/mip/solve.cu +++ b/cpp/src/mip/solve.cu @@ -266,7 +266,7 @@ mip_solution_t solve_mip(optimization_problem_t& op_problem, CUOPT_LOG_INFO("%d implied integers", presolve_result->implied_integer_indices.size()); } if (problem.is_objective_integral()) { CUOPT_LOG_INFO("Objective function is integral"); } - CUOPT_LOG_INFO("Papilo presolve time: %f", presolve_time); + CUOPT_LOG_INFO("Papilo presolve time: %.2f", presolve_time); } if (settings.user_problem_file != "") { CUOPT_LOG_INFO("Writing user problem to file: %s", settings.user_problem_file.c_str()); diff --git a/cpp/src/mip/solver.cu b/cpp/src/mip/solver.cu index b88d39eaa7..2829babc97 100644 --- a/cpp/src/mip/solver.cu +++ b/cpp/src/mip/solver.cu @@ -106,8 +106,9 @@ solution_t mip_solver_t::run_solver() context.problem_ptr->post_process_solution(sol); return sol; } - dm.timer = timer_; - bool presolve_success = dm.run_presolve(timer_.remaining_time()); + dm.timer = timer_; + const bool run_presolve = context.settings.presolve; + bool presolve_success = run_presolve ? dm.run_presolve(timer_.remaining_time()) : true; if (!presolve_success) { CUOPT_LOG_INFO("Problem proven infeasible in presolve"); solution_t sol(*context.problem_ptr); @@ -115,7 +116,7 @@ solution_t mip_solver_t::run_solver() context.problem_ptr->post_process_solution(sol); return sol; } - if (context.problem_ptr->empty) { + if (run_presolve && context.problem_ptr->empty) { CUOPT_LOG_INFO("Problem full reduced in presolve"); solution_t sol(*context.problem_ptr); sol.set_problem_fully_reduced(); @@ -130,7 +131,7 @@ solution_t mip_solver_t::run_solver() } // if the problem was reduced to a LP: run concurrent LP - if (context.problem_ptr->n_integer_vars == 0) { + if (run_presolve && context.problem_ptr->n_integer_vars == 0) { CUOPT_LOG_INFO("Problem reduced to a LP, running concurrent LP"); pdlp_solver_settings_t settings{}; settings.time_limit = timer_.remaining_time(); @@ -174,11 +175,23 @@ solution_t mip_solver_t::run_solver() branch_and_bound_solution.resize(branch_and_bound_problem.num_cols); // Fill in the settings for branch and bound - branch_and_bound_settings.time_limit = timer_.remaining_time(); + branch_and_bound_settings.time_limit = timer_.get_time_limit(); + branch_and_bound_settings.node_limit = context.settings.node_limit; branch_and_bound_settings.print_presolve_stats = false; branch_and_bound_settings.absolute_mip_gap_tol = context.settings.tolerances.absolute_mip_gap; branch_and_bound_settings.relative_mip_gap_tol = context.settings.tolerances.relative_mip_gap; - branch_and_bound_settings.integer_tol = context.settings.tolerances.integrality_tolerance; + branch_and_bound_settings.integer_tol = context.settings.tolerances.integrality_tolerance; + branch_and_bound_settings.max_cut_passes = context.settings.max_cut_passes; + branch_and_bound_settings.mir_cuts = context.settings.mir_cuts; + branch_and_bound_settings.mixed_integer_gomory_cuts = + context.settings.mixed_integer_gomory_cuts; + branch_and_bound_settings.knapsack_cuts = context.settings.knapsack_cuts; + branch_and_bound_settings.strong_chvatal_gomory_cuts = + context.settings.strong_chvatal_gomory_cuts; + branch_and_bound_settings.reduced_cost_strengthening = + context.settings.reduced_cost_strengthening; + branch_and_bound_settings.cut_change_threshold = context.settings.cut_change_threshold; + branch_and_bound_settings.cut_min_orthogonality = context.settings.cut_min_orthogonality; if (context.settings.num_cpu_threads < 0) { branch_and_bound_settings.num_threads = omp_get_max_threads() - 1; @@ -218,7 +231,7 @@ solution_t mip_solver_t::run_solver() // Create the branch and bound object branch_and_bound = std::make_unique>( - branch_and_bound_problem, branch_and_bound_settings); + branch_and_bound_problem, branch_and_bound_settings, timer_.get_tic_start()); context.branch_and_bound_ptr = branch_and_bound.get(); branch_and_bound->set_concurrent_lp_root_solve(true); auto* stats_ptr = &context.stats; diff --git a/cpp/src/utilities/timer.hpp b/cpp/src/utilities/timer.hpp index 1d1a4881e0..b7ab6a63bd 100644 --- a/cpp/src/utilities/timer.hpp +++ b/cpp/src/utilities/timer.hpp @@ -1,6 +1,6 @@ /* clang-format off */ /* - * SPDX-FileCopyrightText: Copyright (c) 2023-2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved. + * SPDX-FileCopyrightText: Copyright (c) 2023-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. * SPDX-License-Identifier: Apache-2.0 */ /* clang-format on */ @@ -55,6 +55,35 @@ class timer_t { double get_time_limit() const noexcept { return time_limit; } + double get_tic_start() const noexcept + { + /** + * Converts a std::chrono::steady_clock::time_point to a struct timeval. + * This is an approximate conversion because steady_clock is relative to an + * unspecified epoch (e.g., system boot time), not the system clock epoch (UTC). + */ + // Get the current time from both clocks at approximately the same instant + std::chrono::system_clock::time_point sys_now = std::chrono::system_clock::now(); + std::chrono::steady_clock::time_point steady_now = std::chrono::steady_clock::now(); + + // Calculate the difference between the given steady_clock time point and the current steady + // time + auto diff_from_now = begin - steady_now; + + // Apply that same difference to the current system clock time point + std::chrono::system_clock::time_point sys_t = sys_now + diff_from_now; + + // Convert the resulting system_clock time point to microseconds since the system epoch + auto us_since_epoch = + std::chrono::duration_cast(sys_t.time_since_epoch()); + + // Populate the timeval struct + double tv_sec = us_since_epoch.count() / 1000000; + double tv_usec = us_since_epoch.count() % 1000000; + + return tv_sec + 1e-6 * tv_usec; + } + private: double time_limit; steady_clock::time_point begin; diff --git a/cpp/tests/mip/CMakeLists.txt b/cpp/tests/mip/CMakeLists.txt index ce47f31444..43fc273dbe 100644 --- a/cpp/tests/mip/CMakeLists.txt +++ b/cpp/tests/mip/CMakeLists.txt @@ -1,5 +1,5 @@ # cmake-format: off -# SPDX-FileCopyrightText: Copyright (c) 2025 NVIDIA CORPORATION & AFFILIATES. All rights reserved. +# SPDX-FileCopyrightText: Copyright (c) 2025-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. # SPDX-License-Identifier: Apache-2.0 # cmake-format: on @@ -26,6 +26,9 @@ ConfigureTest(INCUMBENT_CALLBACK_TEST ConfigureTest(DOC_EXAMPLE_TEST ${CMAKE_CURRENT_SOURCE_DIR}/doc_example_test.cu ) +ConfigureTest(CUTS_TEST + ${CMAKE_CURRENT_SOURCE_DIR}/cuts_test.cu +) ConfigureTest(UNIT_TEST ${CMAKE_CURRENT_SOURCE_DIR}/unit_test.cu ${CMAKE_CURRENT_SOURCE_DIR}/integer_with_real_bounds.cu diff --git a/cpp/tests/mip/cuts_test.cu b/cpp/tests/mip/cuts_test.cu new file mode 100644 index 0000000000..72b9acd47a --- /dev/null +++ b/cpp/tests/mip/cuts_test.cu @@ -0,0 +1,168 @@ +/* clang-format off */ +/* + * SPDX-FileCopyrightText: Copyright (c) 2024-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. + * SPDX-License-Identifier: Apache-2.0 + */ +/* clang-format on */ + +#include "../linear_programming/utilities/pdlp_test_utilities.cuh" +#include "mip_utils.cuh" + +#include +#include +#include +#include + +#include +#include + +#include + +#include +#include +#include +#include +#include +#include + +namespace cuopt::linear_programming::test { + +// Problem data for the mixed integer linear programming problem +mps_parser::mps_data_model_t create_cuts_problem_1() +{ + // Create problem instance + mps_parser::mps_data_model_t problem; + + // Solve the problem + // minimize -7*x1 -2*x2 + // subject to -1*x1 + 2*x2 <= 4 + // 5*x1 + 1*x2 <= 20 + // -2*x1 -2*x2 <= -7 + + // Set up constraint matrix in CSR format + std::vector offsets = {0, 2, 4, 6}; + std::vector indices = {0, 1, 0, 1, 0, 1}; + std::vector coefficients = {-1.0, 2.0, 5.0, 1.0, -2.0, -2.0}; + problem.set_csr_constraint_matrix(coefficients.data(), + coefficients.size(), + indices.data(), + indices.size(), + offsets.data(), + offsets.size()); + + // Set constraint bounds + std::vector lower_bounds = {-std::numeric_limits::infinity(), + -std::numeric_limits::infinity(), + -std::numeric_limits::infinity()}; + std::vector upper_bounds = {4.0, 20.0, -7.0}; + problem.set_constraint_lower_bounds(lower_bounds.data(), lower_bounds.size()); + problem.set_constraint_upper_bounds(upper_bounds.data(), upper_bounds.size()); + + // Set variable bounds + std::vector var_lower_bounds = {0.0, 0.0}; + std::vector var_upper_bounds = {10.0, 10.0}; + problem.set_variable_lower_bounds(var_lower_bounds.data(), var_lower_bounds.size()); + problem.set_variable_upper_bounds(var_upper_bounds.data(), var_upper_bounds.size()); + + // Set objective coefficients (minimize -7*x1 -2*x2) + std::vector objective_coefficients = {-7.0, -2.0}; + problem.set_objective_coefficients(objective_coefficients.data(), objective_coefficients.size()); + + // Set variable types + std::vector variable_types = {'I', 'I'}; + problem.set_variable_types(variable_types); + + return problem; +} + +TEST(cuts, test_cuts_1) +{ + const raft::handle_t handle_{}; + mip_solver_settings_t settings; + constexpr double test_time_limit = 1.; + + // Create the problem + auto problem = create_cuts_problem_1(); + + settings.time_limit = test_time_limit; + settings.max_cut_passes = 1; + mip_solution_t solution = solve_mip(&handle_, problem, settings); + EXPECT_EQ(solution.get_termination_status(), mip_termination_status_t::Optimal); + + double obj_val = solution.get_objective_value(); + // Expected objective value from documentation example is approximately -28 + EXPECT_NEAR(-28, obj_val, 1e-3); + + EXPECT_EQ(solution.get_num_nodes(), 0); +} + +// Problem data for the mixed integer linear programming problem +mps_parser::mps_data_model_t create_cuts_problem_2() +{ + // Create problem instance + mps_parser::mps_data_model_t problem; + + // Solve the problem + // minimize -86*y1 -4*y2 -40*y3 + // subject to 774*y1 + 76*y2 + 42*y3 <= 875 + // 67*y1 + 27*y2 + 53*y3 <= 875 + // y1, y2, y3 in {0, 1} + + // Set up constraint matrix in CSR format + std::vector offsets = {0, 3, 6}; + std::vector indices = {0, 1, 2, 0, 1, 2}; + std::vector coefficients = {774.0, 76.0, 42.0, 67.0, 27.0, 53.0}; + problem.set_csr_constraint_matrix(coefficients.data(), + coefficients.size(), + indices.data(), + indices.size(), + offsets.data(), + offsets.size()); + + // Set constraint bounds + std::vector lower_bounds = {-std::numeric_limits::infinity(), + -std::numeric_limits::infinity()}; + std::vector upper_bounds = {875.0, 875.0}; + problem.set_constraint_lower_bounds(lower_bounds.data(), lower_bounds.size()); + problem.set_constraint_upper_bounds(upper_bounds.data(), upper_bounds.size()); + + // Set variable bounds + std::vector var_lower_bounds = {0.0, 0.0, 0.0}; + std::vector var_upper_bounds = {1.0, 1.0, 1.0}; + problem.set_variable_lower_bounds(var_lower_bounds.data(), var_lower_bounds.size()); + problem.set_variable_upper_bounds(var_upper_bounds.data(), var_upper_bounds.size()); + + // Set objective coefficients (minimize -86*y1 -4*y2 -40*y3) + std::vector objective_coefficients = {-86.0, -4.0, -40.0}; + problem.set_objective_coefficients(objective_coefficients.data(), objective_coefficients.size()); + + // Set variable types + std::vector variable_types = {'I', 'I', 'I'}; + problem.set_variable_types(variable_types); + + return problem; +} + +TEST(cuts, test_cuts_2) +{ + const raft::handle_t handle_{}; + mip_solver_settings_t settings; + constexpr double test_time_limit = 1.; + + // Create the problem + auto problem = create_cuts_problem_2(); + + settings.time_limit = test_time_limit; + settings.max_cut_passes = 10; + settings.presolve = false; + mip_solution_t solution = solve_mip(&handle_, problem, settings); + EXPECT_EQ(solution.get_termination_status(), mip_termination_status_t::Optimal); + + double obj_val = solution.get_objective_value(); + // Expected objective value from documentation example is approximately -126 + EXPECT_NEAR(-126, obj_val, 1e-3); + + EXPECT_EQ(solution.get_num_nodes(), 0); +} + +} // namespace cuopt::linear_programming::test diff --git a/cpp/tests/mip/termination_test.cu b/cpp/tests/mip/termination_test.cu index 849fded55e..3627dd7f40 100644 --- a/cpp/tests/mip/termination_test.cu +++ b/cpp/tests/mip/termination_test.cu @@ -1,6 +1,6 @@ /* clang-format off */ /* - * SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved. + * SPDX-FileCopyrightText: Copyright (c) 2025-2026, NVIDIA CORPORATION & AFFILIATES. All rights reserved. * SPDX-License-Identifier: Apache-2.0 */ /* clang-format on */ @@ -88,7 +88,7 @@ TEST(termination_status, optimality_test) auto [termination_status, obj_val, lb] = test_mps_file("mip/bb_optimality.mps", default_time_limit, false); EXPECT_EQ(termination_status, mip_termination_status_t::Optimal); - EXPECT_EQ(obj_val, 2); + EXPECT_NEAR(obj_val, 2, 1e-6); } // Ensure the lower bound on maximization problems when BB times out has the right sign diff --git a/python/cuopt/cuopt/tests/linear_programming/test_python_API.py b/python/cuopt/cuopt/tests/linear_programming/test_python_API.py index 71befa512f..156c7ed69f 100644 --- a/python/cuopt/cuopt/tests/linear_programming/test_python_API.py +++ b/python/cuopt/cuopt/tests/linear_programming/test_python_API.py @@ -395,9 +395,10 @@ def set_solution( x_val = sol["solution"][0] y_val = sol["solution"][1] cost = sol["cost"] - assert 2 * x_val + 4 * y_val >= 230 - assert 3 * x_val + 2 * y_val <= 190 - assert 5 * x_val + 3 * y_val == cost + tol = 1e-6 + assert 2 * x_val + 4 * y_val >= 230 - tol + assert 3 * x_val + 2 * y_val <= 190 + tol + assert abs(5 * x_val + 3 * y_val - cost) < tol def test_incumbent_get_solutions(): diff --git a/python/libcuopt/pyproject.toml b/python/libcuopt/pyproject.toml index e2c1142523..4ddc9ca756 100644 --- a/python/libcuopt/pyproject.toml +++ b/python/libcuopt/pyproject.toml @@ -53,7 +53,6 @@ libcuopt = "libcuopt" select = [ "distro-too-large-compressed", ] - max_allowed_size_compressed = '660M' [project.scripts] From 9c3c72785f5c19b20df55ba585d736223615791b Mon Sep 17 00:00:00 2001 From: Chris Maes Date: Thu, 5 Feb 2026 17:58:28 -0800 Subject: [PATCH 3/3] Adjust basis repair for super basic variables. Repeat root solve if cuts make it infeasible (#831) Fixes an issue on neos-4413714-turia where basis repair was called during dual push, and the slack needed was superbasic. Fixes an issue where cbs-cta was incorrectly classified as infeasible after cuts were added. Authors: - Chris Maes (https://github.com/chris-maes) Approvers: - Alice Boucher (https://github.com/aliceb-nv) URL: https://github.com/NVIDIA/cuopt/pull/831 --- cpp/src/dual_simplex/basis_solves.cpp | 39 ++++++--- cpp/src/dual_simplex/basis_solves.hpp | 1 + cpp/src/dual_simplex/basis_updates.cpp | 13 ++- cpp/src/dual_simplex/branch_and_bound.cpp | 28 +++++-- cpp/src/dual_simplex/crossover.cpp | 98 ++++++++++++++--------- cpp/src/dual_simplex/primal.cpp | 1 + 6 files changed, 123 insertions(+), 57 deletions(-) diff --git a/cpp/src/dual_simplex/basis_solves.cpp b/cpp/src/dual_simplex/basis_solves.cpp index 2ee5e1d01f..3b26b2a8b9 100644 --- a/cpp/src/dual_simplex/basis_solves.cpp +++ b/cpp/src/dual_simplex/basis_solves.cpp @@ -619,12 +619,13 @@ i_t basis_repair(const csc_matrix_t& A, const std::vector& slacks_needed, std::vector& basis_list, std::vector& nonbasic_list, + std::vector& superbasic_list, std::vector& vstatus) { const i_t m = A.m; const i_t n = A.n; assert(basis_list.size() == m); - assert(nonbasic_list.size() == n - m); + assert(nonbasic_list.size() + superbasic_list.size() == n - m); // Create slack_map std::vector slack_map(m); // slack_map[i] = j if column j is e_i @@ -649,6 +650,13 @@ i_t basis_repair(const csc_matrix_t& A, for (i_t k = 0; k < num_nonbasic; ++k) { nonbasic_map[nonbasic_list[k]] = k; } + // Create a superbasic_map + std::vector superbasic_map( + n, -1); // superbasic_map[j] = p if superbasic[p] = j, -1 if j is basic/nonbasic + const i_t num_superbasic = superbasic_list.size(); + for (i_t k = 0; k < num_superbasic; ++k) { + superbasic_map[superbasic_list[k]] = k; + } const i_t columns_to_replace = deficient.size(); for (i_t k = 0; k < columns_to_replace; ++k) { @@ -656,19 +664,25 @@ i_t basis_repair(const csc_matrix_t& A, const i_t replace_i = slacks_needed[k]; const i_t replace_j = slack_map[replace_i]; basis_list[deficient[k]] = replace_j; - assert(nonbasic_map[replace_j] != -1); - nonbasic_list[nonbasic_map[replace_j]] = bad_j; - vstatus[replace_j] = variable_status_t::BASIC; - // This is the main issue. What value should bad_j take on. - if (lower[bad_j] == -inf && upper[bad_j] == inf) { - vstatus[bad_j] = variable_status_t::NONBASIC_FREE; - } else if (lower[bad_j] > -inf) { - vstatus[bad_j] = variable_status_t::NONBASIC_LOWER; - } else if (upper[bad_j] < inf) { - vstatus[bad_j] = variable_status_t::NONBASIC_UPPER; + if (nonbasic_map[replace_j] != -1) { + nonbasic_list[nonbasic_map[replace_j]] = bad_j; + // This is the main issue. What value should bad_j take on. + if (lower[bad_j] == -inf && upper[bad_j] == inf) { + vstatus[bad_j] = variable_status_t::NONBASIC_FREE; + } else if (lower[bad_j] > -inf) { + vstatus[bad_j] = variable_status_t::NONBASIC_LOWER; + } else if (upper[bad_j] < inf) { + vstatus[bad_j] = variable_status_t::NONBASIC_UPPER; + } else { + assert(1 == 0); + } + } else if (superbasic_map[replace_j] != -1) { + superbasic_list[superbasic_map[replace_j]] = bad_j; + vstatus[bad_j] = variable_status_t::SUPERBASIC; } else { - assert(1 == 0); + assert(nonbasic_map[replace_j] != -1 || superbasic_map[replace_j] != -1); } + vstatus[replace_j] = variable_status_t::BASIC; } return 0; @@ -865,6 +879,7 @@ template int basis_repair(const csc_matrix_t& A, const std::vector& slacks_needed, std::vector& basis_list, std::vector& nonbasic_list, + std::vector& superbasic_list, std::vector& vstatus); template int form_b(const csc_matrix_t& A, diff --git a/cpp/src/dual_simplex/basis_solves.hpp b/cpp/src/dual_simplex/basis_solves.hpp index 295bedccdd..59b4725e42 100644 --- a/cpp/src/dual_simplex/basis_solves.hpp +++ b/cpp/src/dual_simplex/basis_solves.hpp @@ -48,6 +48,7 @@ i_t basis_repair(const csc_matrix_t& A, const std::vector& slacks_needed, std::vector& basis_list, std::vector& nonbasic_list, + std::vector& superbasic_list, std::vector& vstatus); // Form the basis matrix B = A(:, basic_list) diff --git a/cpp/src/dual_simplex/basis_updates.cpp b/cpp/src/dual_simplex/basis_updates.cpp index 1d94f41c7f..dd262622c2 100644 --- a/cpp/src/dual_simplex/basis_updates.cpp +++ b/cpp/src/dual_simplex/basis_updates.cpp @@ -2265,6 +2265,7 @@ int basis_update_mpf_t::refactor_basis( { std::vector deficient; std::vector slacks_needed; + std::vector superbasic_list; // Empty superbasic list if (L0_.m != A.m) { resize(A.m); } std::vector q; @@ -2281,8 +2282,16 @@ int basis_update_mpf_t::refactor_basis( if (status == CONCURRENT_HALT_RETURN) { return CONCURRENT_HALT_RETURN; } if (status == -1) { settings.log.debug("Initial factorization failed\n"); - basis_repair( - A, settings, lower, upper, deficient, slacks_needed, basic_list, nonbasic_list, vstatus); + basis_repair(A, + settings, + lower, + upper, + deficient, + slacks_needed, + basic_list, + nonbasic_list, + superbasic_list, + vstatus); #ifdef CHECK_BASIS_REPAIR const i_t m = A.m; diff --git a/cpp/src/dual_simplex/branch_and_bound.cpp b/cpp/src/dual_simplex/branch_and_bound.cpp index 7f6e2c1921..d354159046 100644 --- a/cpp/src/dual_simplex/branch_and_bound.cpp +++ b/cpp/src/dual_simplex/branch_and_bound.cpp @@ -1991,7 +1991,9 @@ mip_status_t branch_and_bound_t::solve(mip_solution_t& solut root_relax_soln_, iter, edge_norms_); - f_t dual_phase2_time = toc(dual_phase2_start_time); + exploration_stats_.total_lp_iters += iter; + root_objective_ = compute_objective(original_lp_, root_relax_soln_.x); + f_t dual_phase2_time = toc(dual_phase2_start_time); if (dual_phase2_time > 1.0) { settings_.log.debug("Dual phase2 time %.2f seconds\n", dual_phase2_time); } @@ -2002,11 +2004,27 @@ mip_status_t branch_and_bound_t::solve(mip_solution_t& solut } if (cut_status != dual::status_t::OPTIMAL) { - settings_.log.printf("Cut status %s\n", dual::status_to_string(cut_status).c_str()); - return mip_status_t::NUMERICAL; + settings_.log.printf("Numerical issue at root node. Resolving from scratch\n"); + lp_status_t scratch_status = + solve_linear_program_with_advanced_basis(original_lp_, + exploration_stats_.start_time, + lp_settings, + root_relax_soln_, + basis_update, + basic_list, + nonbasic_list, + root_vstatus_, + edge_norms_); + if (scratch_status == lp_status_t::OPTIMAL) { + // We recovered + cut_status = convert_lp_status_to_dual_status(scratch_status); + exploration_stats_.total_lp_iters += root_relax_soln_.iterations; + root_objective_ = compute_objective(original_lp_, root_relax_soln_.x); + } else { + settings_.log.printf("Cut status %s\n", dual::status_to_string(cut_status).c_str()); + return mip_status_t::NUMERICAL; + } } - exploration_stats_.total_lp_iters += root_relax_soln_.iterations; - root_objective_ = compute_objective(original_lp_, root_relax_soln_.x); local_lower_bounds_.assign(settings_.num_bfs_workers, root_objective_); diff --git a/cpp/src/dual_simplex/crossover.cpp b/cpp/src/dual_simplex/crossover.cpp index 81d5ec1e6d..afc8c66744 100644 --- a/cpp/src/dual_simplex/crossover.cpp +++ b/cpp/src/dual_simplex/crossover.cpp @@ -516,6 +516,7 @@ i_t dual_push(const lp_problem_t& lp, slacks_needed, basic_list, nonbasic_list, + superbasic_list, vstatus); rank = factorize_basis(lp.A, settings, basic_list, L, U, p, pinv, q, deficient, slacks_needed); @@ -585,6 +586,55 @@ f_t primal_residual(const lp_problem_t& lp, const lp_solution_t(primal_residual); } +template +void find_primal_superbasic_variables(const lp_problem_t& lp, + const simplex_solver_settings_t& settings, + const lp_solution_t& initial_solution, + lp_solution_t& solution, + std::vector& vstatus, + std::vector& nonbasic_list, + std::vector& superbasic_list) +{ + const i_t n = lp.num_cols; + const f_t fixed_tolerance = settings.fixed_tol; + constexpr f_t basis_threshold = 1e-6; + nonbasic_list.clear(); + superbasic_list.clear(); + + for (i_t j = 0; j < n; ++j) { + if (vstatus[j] != variable_status_t::BASIC) { + const f_t lower_infeas = lp.lower[j] - initial_solution.x[j]; + const f_t lower_bound_slack = initial_solution.x[j] - lp.lower[j]; + const f_t upper_infeas = initial_solution.x[j] - lp.upper[j]; + const f_t upper_bound_slack = lp.upper[j] - initial_solution.x[j]; + if (std::abs(lp.lower[j] - lp.upper[j]) < fixed_tolerance) { + vstatus[j] = variable_status_t::NONBASIC_FIXED; + nonbasic_list.push_back(j); + } else if (lower_infeas > 0 && lp.lower[j] > -inf) { + vstatus[j] = variable_status_t::NONBASIC_LOWER; + solution.x[j] = lp.lower[j]; + nonbasic_list.push_back(j); + } else if (upper_infeas > 0 && lp.upper[j] < inf) { + vstatus[j] = variable_status_t::NONBASIC_UPPER; + solution.x[j] = lp.upper[j]; + nonbasic_list.push_back(j); + } else if (lower_bound_slack < basis_threshold && lp.lower[j] > -inf) { + vstatus[j] = variable_status_t::NONBASIC_LOWER; + nonbasic_list.push_back(j); + } else if (upper_bound_slack < basis_threshold && lp.upper[j] < inf) { + vstatus[j] = variable_status_t::NONBASIC_UPPER; + nonbasic_list.push_back(j); + } else if (lp.lower[j] == -inf && lp.upper[j] == inf) { + vstatus[j] = variable_status_t::NONBASIC_FREE; + nonbasic_list.push_back(j); + } else { + vstatus[j] = variable_status_t::SUPERBASIC; + superbasic_list.push_back(j); + } + } + } +} + template f_t primal_ratio_test(const lp_problem_t& lp, const simplex_solver_settings_t& settings, @@ -784,6 +834,7 @@ i_t primal_push(const lp_problem_t& lp, } basic_list[basic_leaving_index] = s; nonbasic_list.push_back(leaving_index); + superbasic_list.pop_back(); // Remove superbasic variable // Refactor or Update bool should_refactor = ft.num_updates() > 100; @@ -820,7 +871,11 @@ i_t primal_push(const lp_problem_t& lp, slacks_needed, basic_list, nonbasic_list, + superbasic_list, vstatus); + // We need to be careful. As basis_repair may have changed the superbasic list + find_primal_superbasic_variables( + lp, settings, solution, solution, vstatus, nonbasic_list, superbasic_list); rank = factorize_basis(lp.A, settings, basic_list, L, U, p, pinv, q, deficient, slacks_needed); if (rank == CONCURRENT_HALT_RETURN) { @@ -844,11 +899,9 @@ i_t primal_push(const lp_problem_t& lp, vstatus[s] = variable_status_t::NONBASIC_UPPER; nonbasic_list.push_back(s); } + superbasic_list.pop_back(); // Remove superbasic variable } - // Remove superbasic variable - superbasic_list.pop_back(); - num_pushes++; if (num_pushes % settings.iteration_log_frequency == 0 || toc(last_print_time) > 10.0 || superbasic_list.size() == 0) { @@ -1178,6 +1231,7 @@ crossover_status_t crossover(const lp_problem_t& lp, slacks_needed, basic_list, nonbasic_list, + superbasic_list, vstatus); rank = factorize_basis(lp.A, settings, basic_list, L, U, p, pinv, q, deficient, slacks_needed); if (rank == CONCURRENT_HALT_RETURN) { @@ -1210,41 +1264,8 @@ crossover_status_t crossover(const lp_problem_t& lp, settings.log.debug("nonbasic list size %ld n - m %d\n", nonbasic_list.size(), n - m); print_crossover_info(lp, settings, vstatus, solution, "Dual push complete"); - nonbasic_list.clear(); - superbasic_list.clear(); - - for (i_t j = 0; j < n; ++j) { - if (vstatus[j] != variable_status_t::BASIC) { - const f_t lower_infeas = lp.lower[j] - initial_solution.x[j]; - const f_t lower_bound_slack = initial_solution.x[j] - lp.lower[j]; - const f_t upper_infeas = initial_solution.x[j] - lp.upper[j]; - const f_t upper_bound_slack = lp.upper[j] - initial_solution.x[j]; - if (std::abs(lp.lower[j] - lp.upper[j]) < fixed_tolerance) { - vstatus[j] = variable_status_t::NONBASIC_FIXED; - nonbasic_list.push_back(j); - } else if (lower_infeas > 0 && lp.lower[j] > -inf) { - vstatus[j] = variable_status_t::NONBASIC_LOWER; - solution.x[j] = lp.lower[j]; - nonbasic_list.push_back(j); - } else if (upper_infeas > 0 && lp.upper[j] < inf) { - vstatus[j] = variable_status_t::NONBASIC_UPPER; - solution.x[j] = lp.upper[j]; - nonbasic_list.push_back(j); - } else if (lower_bound_slack < basis_threshold && lp.lower[j] > -inf) { - vstatus[j] = variable_status_t::NONBASIC_LOWER; - nonbasic_list.push_back(j); - } else if (upper_bound_slack < basis_threshold && lp.upper[j] < inf) { - vstatus[j] = variable_status_t::NONBASIC_UPPER; - nonbasic_list.push_back(j); - } else if (lp.lower[j] == -inf && lp.upper[j] == inf) { - vstatus[j] = variable_status_t::NONBASIC_FREE; - nonbasic_list.push_back(j); - } else { - vstatus[j] = variable_status_t::SUPERBASIC; - superbasic_list.push_back(j); - } - } - } + find_primal_superbasic_variables( + lp, settings, initial_solution, solution, vstatus, nonbasic_list, superbasic_list); if (superbasic_list.size() > 0) { std::vector save_x = solution.x; @@ -1381,6 +1402,7 @@ crossover_status_t crossover(const lp_problem_t& lp, slacks_needed, basic_list, nonbasic_list, + superbasic_list, vstatus); rank = factorize_basis(lp.A, settings, basic_list, L, U, p, pinv, q, deficient, slacks_needed); diff --git a/cpp/src/dual_simplex/primal.cpp b/cpp/src/dual_simplex/primal.cpp index 38cddc0e24..98f5f4193b 100644 --- a/cpp/src/dual_simplex/primal.cpp +++ b/cpp/src/dual_simplex/primal.cpp @@ -308,6 +308,7 @@ primal::status_t primal_phase2(i_t phase, slacks_needed, basic_list, nonbasic_list, + superbasic_list, vstatus); rank = factorize_basis(lp.A, settings, basic_list, L, U, p, pinv, q, deficient, slacks_needed); if (rank == CONCURRENT_HALT_RETURN) {