Frank Vega
Information Physics Institute, 840 W 67th St, Hialeah, FL 33012, USA
[email protected]

Problem Statement

Given an undirected graph G=(V,E)G = (V, E) , the goal is to find a dominating set SVS \subseteq V , where a set SS is dominating if every vertex in VV is either in SS or adjacent to a vertex in SS . We aim to design an algorithm that produces a dominating set SS such that S2OPT|S| \leq 2 \cdot |OPT| , where OPTOPT is the size of a minimum dominating set in GG , thus achieving a 2-approximation.


Algorithm Description

Consider the algorithm implemented in Python:

import networkx as nx

def find_dominating_set(graph):
    """
    Approximate minimum dominating set for an undirected graph by transforming it into a bipartite graph.

    Args:
        graph (nx.Graph): A NetworkX Graph object representing the input graph.

    Returns:
        set: A set of vertex indices representing the approximate minimum dominating set.
             Returns an empty set if the graph is empty or has no edges.
    """
    # Subroutine to compute a dominating set in a bipartite component, used to find a dominating set in the original graph
    def find_dominating_set_via_bipartite_proxy(G):
        # Initialize an empty set to store the dominating set for this bipartite component
        dominating_set = set()
        # Track which vertices in the bipartite graph are dominated
        dominated = {v: False for v in G.nodes()}
        # Sort vertices by degree (ascending) to prioritize high-degree nodes for greedy selection
        undominated = sorted(list(G.nodes()), key=lambda x: G.degree(x))

        # Continue processing until all vertices are dominated
        while undominated:
            # Pop the next vertex to process (starting with highest degree)
            v = undominated.pop()
            # Check if the vertex is not yet dominated
            if not dominated[v]:
                # Initialize the best vertex to add as the current vertex
                best_vertex = v
                # Initialize the count of undominated vertices covered by the best vertex
                best_undominated_count = -1

                # Consider the current vertex and its neighbors as candidates
                for neighbor in list(G.neighbors(v)) + [v]:
                    # Count how many undominated vertices this candidate covers
                    undominated_neighbors_count = 0
                    for u in list(G.neighbors(neighbor)) + [neighbor]:
                        if not dominated[u]:
                            undominated_neighbors_count += 1

                    # Update the best vertex if this candidate covers more undominated vertices
                    if undominated_neighbors_count > best_undominated_count:
                        best_undominated_count = undominated_neighbors_count
                        best_vertex = neighbor

                # Add the best vertex to the dominating set for this component
                dominating_set.add(best_vertex)

                # Mark the neighbors of the best vertex as dominated
                for neighbor in G.neighbors(best_vertex):
                    dominated[neighbor] = True
                    # Mark the mirror vertex (i, 1-k) as dominated to reflect domination in the original graph
                    mirror_neighbor = (neighbor[0], 1 - neighbor[1])
                    dominated[mirror_neighbor] = True

        # Return the dominating set for this bipartite component
        return dominating_set

    # Validate that the input is a NetworkX Graph object
    if not isinstance(graph, nx.Graph):
        raise ValueError("Input must be an undirected NetworkX Graph.")

    # Handle edge cases: return an empty set if the graph has no nodes or no edges
    if graph.number_of_nodes() == 0 or graph.number_of_edges() == 0:
        return set()

    # Initialize the dominating set with all isolated nodes, as they must be included to dominate themselves
    approximate_dominating_set = set(nx.isolates(graph))
    # Remove isolated nodes from the graph to process the remaining components
    graph.remove_nodes_from(approximate_dominating_set)

    # If the graph is empty after removing isolated nodes, return the set of isolated nodes
    if graph.number_of_nodes() == 0:
        return approximate_dominating_set

    # Initialize an empty bipartite graph to transform the remaining graph
    bipartite_graph = nx.Graph()

    # Construct the bipartite graph B
    for i in graph.nodes():
        # Add an edge between mirror nodes (i, 0) and (i, 1) for each vertex i
        bipartite_graph.add_edge((i, 0), (i, 1))
        # Add edges reflecting adjacency in the original graph: (i, 0) to (j, 1) for each neighbor j
        for j in graph.neighbors(i):
            bipartite_graph.add_edge((i, 0), (j, 1))

    # Process each connected component in the bipartite graph
    for component in nx.connected_components(bipartite_graph):
        # Extract the subgraph for the current connected component
        bipartite_subgraph = bipartite_graph.subgraph(component)

        # Compute the dominating set for this component using the subroutine
        tuple_nodes = find_dominating_set_via_bipartite_proxy(bipartite_subgraph)

        # Extract the original node indices from the tuple nodes (i, k) and add them to the dominating set
        approximate_dominating_set.update({tuple_node[0] for tuple_node in tuple_nodes})

    # Return the final dominating set for the original graph
    return approximate_dominating_set

The algorithm transforms the problem into a bipartite graph setting and uses a greedy approach. Here are the steps:

  1. Handle Isolated Nodes:

    • Identify all isolated nodes in GG (vertices with degree 0).
    • Add these nodes to the dominating set SS , as they must be included to dominate themselves.
  2. Construct a Bipartite Graph BB :

    • For the remaining graph (after removing isolated nodes), construct a bipartite graph BB with:
      • Vertex Set: Two partitions, where each vertex iVi \in V is duplicated as (i,0)(i, 0) and (i,1)(i, 1) .
      • Edge Set:
      • An edge (i,0)(i,1)(i, 0) - (i, 1) for each iVi \in V .
      • For each edge (i,j)E(i, j) \in E in GG , add edges (i,0)(j,1)(i, 0) - (j, 1) and (j,0)(i,1)(j, 0) - (i, 1) in BB .
  3. Greedy Dominating Set in BB :

    • Run a greedy algorithm on BB to compute a dominating set DBD_B :
      • While there are undominated vertices in BB , select the vertex that dominates the maximum number of currently undominated vertices and add it to DBD_B .
  4. Map Back to GG :

    • Define the dominating set SS for GG as:
      • S={iV(i,0)DB or (i,1)DB}S = \{ i \in V \mid (i, 0) \in D_B \text{ or } (i, 1) \in D_B \} .
    • Include the isolated nodes identified in Step 1.

Correctness of the Algorithm

Let’s verify that SS is a dominating set for GG :

  • Isolated Nodes:

    • All isolated nodes are explicitly added to SS , so they are dominated by themselves.
  • Non-Isolated Nodes:

    • Consider any vertex iVi \in V in the non-isolated part of GG :
    • Case 1: iSi \in S :
      • If (i,0)(i, 0) or (i,1)(i, 1) is in DBD_B , then iSi \in S , and ii is dominated by itself.
    • Case 2: iSi \notin S :
      • If neither (i,0)(i, 0) nor (i,1)(i, 1) is in DBD_B , then since DBD_B dominates all vertices in BB , (i,0)(i, 0) must be adjacent to some (j,1)DB(j, 1) \in D_B .
      • In BB , (i,0)(j,1)(i, 0) - (j, 1) exists if (i,j)E(i, j) \in E in GG .
      • Thus, jSj \in S (because (j,1)DB(j, 1) \in D_B ), and ii is adjacent to jj in GG .
  • Conclusion:

    • Every vertex in GG is either in SS or has a neighbor in SS , so SS is a dominating set.

Approximation Analysis Using DuD_u

To prove that S2OPT|S| \leq 2 \cdot |OPT| , we associate each vertex in the optimal dominating set OPTOPT of GG with vertices in SS , ensuring that each vertex in OPTOPT is “responsible” for at most two vertices in SS .

Definitions

  • Let OPTOPT be a minimum dominating set of GG .
  • For each vertex uOPTu \in OPT , define DuSD_u \subseteq S as the set of vertices in SS that are charged to uu based on how the greedy algorithm selects vertices in BB to dominate vertices related to uu .

Key Idea

The DuD_u analysis shows that the greedy algorithm selects at most two vertices per vertex in OPTOPT due to the graph’s structure. Here, we mirror this by analyzing the bipartite graph BB and the mapping back to GG , showing that each uOPTu \in OPT contributes to at most two vertices being added to SS .

Analysis Steps

  1. Role of OPTOPT in GG :

    • Each uOPTu \in OPT dominates itself and its neighbors in GG .
    • In BB , the vertices (u,0)(u, 0) and (u,1)(u, 1) correspond to uu , and we need to ensure they (and their neighbors) are dominated by DBD_B .
  2. Greedy Selection in BB :

    • The greedy algorithm selects a vertex w=(j,k)w = (j, k) (where k=0k = 0 or 11 ) in BB to maximize the number of undominated vertices covered.
    • When ww is added to DBD_B , jj is added to SS .
  3. Defining DuD_u :

    • For each uOPTu \in OPT , DuD_u includes vertices jSj \in S such that the selection of (j,0)(j, 0) or (j,1)(j, 1) in DBD_B helps dominate (u,0)(u, 0) or (u,1)(u, 1) .
    • We charge jj to uu if:
      • j=uj = u (i.e., (u,0)(u, 0) or (u,1)(u, 1) is selected), or
      • jj is a neighbor of uu in GG , and selecting (j,k)(j, k) dominates (u,0)(u, 0) or (u,1)(u, 1) via an edge in BB .
  4. Bounding Du|D_u| :

    • Case 1: uSu \in S :
      • If (u,0)(u, 0) or (u,1)(u, 1) is selected in DBD_B , then uSu \in S .
      • Selecting (u,1)(u, 1) dominates (u,0)(u, 0) (via (u,0)(u,1)(u, 0) - (u, 1) ), and vice versa.
      • At most one vertex ( uu ) is added to SS for uu , so Du1|D_u| \leq 1 in this case.
    • Case 2: uSu \notin S :
      • Neither (u,0)(u, 0) nor (u,1)(u, 1) is in DBD_B .
      • (u,0)(u, 0) must be dominated by some (j,1)DB(j, 1) \in D_B , where (u,j)E(u, j) \in E in GG , so jSj \in S .
      • (u,1)(u, 1) must be dominated by some (k,0)DB(k, 0) \in D_B , where (u,k)E(u, k) \in E in GG , so kSk \in S .
      • Thus, Du={j,k}D_u = \{ j, k \} , and Du2|D_u| \leq 2 (if j=kj = k , then Du=1|D_u| = 1 ).
    • Greedy Optimization:
      • The greedy algorithm often selects a single vertex that dominates both (u,0)(u, 0) and (u,1)(u, 1) indirectly through neighbors, but in the worst case, two selections suffice.
  5. Total Size of SS :

    • Each vertex jSj \in S corresponds to at least one selection (j,k)DB(j, k) \in D_B .
    • Since each uOPTu \in OPT has Du2|D_u| \leq 2 , the total number of vertices in SS is:
      SuOPTDu2OPT|S| \leq \sum_{u \in OPT} |D_u| \leq 2 \cdot |OPT|
    • This accounts for all selections each of which corresponds to a distinct
      uOPT.u \in OPT.

Conclusion

  • Each uOPTu \in OPT is associated with at most two vertices in SS via the DuD_u charging scheme.
  • Thus, S2OPT|S| \leq 2 \cdot |OPT| , proving the 2-approximation.

Conclusion

The algorithm computes a dominating set SS for GG by:

  • Adding all isolated nodes to SS .
  • Constructing a bipartite graph BB with vertices (i,0)(i, 0) and (i,1)(i, 1) for each iVi \in V and edges reflecting GG ’s structure.
  • Using a greedy algorithm to find a dominating set DBD_B in BB .
  • Mapping back to S={i(i,0)DB or (i,1)DB}S = \{ i \mid (i, 0) \in D_B \text{ or } (i, 1) \in D_B \} .

SS is a dominating set because every vertex in GG is either in SS or adjacent to a vertex in SS . By adapting the DuD_u analysis from chordal graphs, we define DuD_u for each uOPTu \in OPT as the vertices in SS charged to uu , showing Du2|D_u| \leq 2 . This ensures S2OPT|S| \leq 2 \cdot |OPT| , achieving a 2-approximation ratio for the dominating set problem in general undirected graphs.

Runtime Analysis of the Algorithm

To analyze the runtime of the find_dominating_set algorithm, we need to examine its key components and determine the time complexity of each step. The algorithm processes a graph to find a dominating set using a bipartite graph construction and a greedy subroutine. Below, we break down the analysis into distinct steps, assuming the input graph GG has nn nodes and mm edges.


Step 1: Handling Isolated Nodes

The algorithm begins by identifying and handling isolated nodes (nodes with no edges) in the graph.

  • Identify Isolated Nodes: Using nx.isolates(graph) from the NetworkX library, isolated nodes are found by checking the degree of each node. This takes O(n)O(n) time, as there are nn nodes to examine.
  • Remove Isolated Nodes: Removing a node in NetworkX is an O(1)O(1) operation per node. If there are kk isolated nodes (where knk \leq n ), the total time to remove them is O(k)O(k) , which is bounded by O(n)O(n) .

Time Complexity for Step 1: O(n)O(n)


Step 2: Constructing the Bipartite Graph

Next, the algorithm constructs a bipartite graph BB based on the remaining graph after isolated nodes are removed.

  • Add Mirror Edges: For each node ii in GG , two nodes (i,0)(i, 0) and (i,1)(i, 1) are created in BB , connected by an edge. With nn nodes, and each edge addition being O(1)O(1) , this step takes O(n)O(n) time.
  • Add Adjacency Edges: For each edge (i,j)(i, j) in GG , edges (i,0)(j,1)(i, 0) - (j, 1) and (j,0)(i,1)(j, 0) - (i, 1) are added to BB . Since GG is undirected, each edge is processed once, and adding two edges per original edge takes O(1)O(1) time. With mm edges, this is O(m)O(m) .

Time Complexity for Step 2: O(n+m)O(n + m)


Step 3: Finding Connected Components

The algorithm identifies connected components in the bipartite graph BB .

  • Compute Connected Components: Using nx.connected_components, this operation runs in O(n+m)O(n' + m') time, where nn' and mm' are the number of nodes and edges in BB . In BB , there are n=2nn' = 2n nodes (two per node in GG ) and m=n+2mm' = n + 2m edges ( nn mirror edges plus 2m2m adjacency edges). Thus, the time is O(2n+(n+2m))=O(n+m)O(2n + (n + 2m)) = O(n + m) .

Time Complexity for Step 3: O(n+m)O(n + m)


Step 4: Computing Dominating Sets for Each Component

For each connected component in BB , the subroutine find_dominating_set_via_bipartite_proxy computes a dominating set. We analyze this subroutine for a single component with ncn_c nodes and mcm_c edges, then scale to all components.

Subroutine Analysis: find_dominating_set_via_bipartite_proxy

  • Initialization:

    • Create a dominated dictionary: O(nc)O(n_c) .
    • Sort nodes by degree: O(nclognc)O(n_c \log n_c) .
  • Main Loop:

    • The loop continues until all nodes are dominated, running up to O(nc)O(n_c) iterations in the worst case.
    • For each iteration:
    • Select a node vv and evaluate it and its neighbors (its closed neighborhood) to find the vertex that dominates the most undominated nodes.
    • For a node vv with degree dvd_v , there are dv+1d_v + 1 candidates (including vv ).
    • For each candidate, count undominated nodes in its closed neighborhood, taking O(dcandidate)O(d_{\text{candidate}}) time per candidate.
    • Total time per iteration is O(dv+uN(v)du)O(d_v + \sum_{u \in N(v)} d_u) , which is the sum of degrees in the closed neighborhood of vv .
    • After selecting the best vertex, mark its neighbors and their mirrors as dominated, taking O(dbest)O(d_{\text{best}}) time.
  • Subroutine Total:

    • Sorting: O(nclognc)O(n_c \log n_c) .
    • Loop: Across all iterations, each node is processed once, and the total work is proportional to the sum of degrees, which is O(mc)O(m_c) .
    • Total for one component: O(nclognc+mc)O(n_c \log n_c + m_c) .

Across All Components

  • The components are disjoint, with nc=2n\sum n_c = 2n and mc=n+2m\sum m_c = n + 2m .
  • The total time is O((nclognc+mc))O(\sum (n_c \log n_c + m_c)) .
  • The nclognc\sum n_c \log n_c term is maximized when all nodes are in one component, yielding O(2nlog2n)=O(nlogn)O(2n \log 2n) = O(n \log n) .
  • The mc=O(n+m)\sum m_c = O(n + m) .

Time Complexity for Step 4: O(nlogn+m)O(n \log n + m)


Overall Time Complexity

Combining all steps:

  • Step 1: O(n)O(n)
  • Step 2: O(n+m)O(n + m)
  • Step 3: O(n+m)O(n + m)
  • Step 4: O(nlogn+m)O(n \log n + m)

The dominant term is O(nlogn+m)O(n \log n + m) , so the overall time complexity of the find_dominating_set algorithm is O(nlogn+m)O(n \log n + m)


Space Complexity

  • Bipartite Graph: O(n+m)O(n + m) , with 2n2n nodes and n+2mn + 2m edges.
  • Auxiliary Data Structures: The dominated dictionary and undominated list use O(n)O(n) space.

Space Complexity: O(n+m)O(n + m)


Conclusion

The find_dominating_set algorithm runs in O(nlogn+m)O(n \log n + m) time and uses O(n+m)O(n + m) space, where nn is the number of nodes and mm is the number of edges in the input graph. The bottleneck arises from sorting nodes by degree in the subroutine, contributing the nlognn \log n factor. This complexity makes the algorithm efficient and scalable for large graphs, especially given its 2-approximation guarantee for the dominating set problem.

Experimental Results

Methodology and Experimental Setup

To evaluate our algorithm rigorously, we use the benchmark instances from the Second DIMACS Implementation Challenge [Johnson1996]. These instances are widely recognized in the computational graph theory community for their diversity and hardness, making them ideal for testing algorithms on the minimum dominating set problem.

The experiments were conducted on a system with the following specifications:

  • Processor: 11th Gen Intel® Core™ i7-1165G7 (2.80 GHz, up to 4.70 GHz with Turbo Boost)
  • Memory: 32 GB DDR4 RAM
  • Operating System: Windows 10 Pro (64-bit)

Our algorithm was implemented using Baldor: Approximate Minimum Dominating Set Solver (v0.1.3) [Vega25], a custom implementation designed to achieve a 2-approximation guarantee for the minimum dominating set problem. As a baseline for comparison, we employed the weighted dominating set approximation algorithm provided by NetworkX [Vazirani2001], which guarantees a solution within a logarithmic approximation ratio of logVOPT\log |V| \cdot OPT , where OPTOPT is the size of the optimal dominating set and V|V| is the number of vertices in the graph.

Each algorithm was run on the same set of DIMACS instances, and the results were recorded to ensure a fair comparison.

Performance Metrics

We evaluate the performance of our algorithm using the following metrics:

  1. Runtime (milliseconds): The total computation time required to compute the dominating set, measured in milliseconds. This metric reflects the algorithm's efficiency and scalability on graphs of varying sizes.
  2. Approximation Quality: To quantify the quality of the solutions produced by our algorithm, we compute the upper bound on the approximation ratio, defined as:
logVOPTBOPTW \log |V| \cdot \frac{|OPT_B|}{|OPT_W|}

where:

  • OPTB|OPT_B| : The size of the dominating set produced by our algorithm (Baldor).
  • OPTW|OPT_W| : The size of the dominating set produced by the NetworkX baseline.
  • V|V| : The number of vertices in the graph.

Given the theoretical guarantees, NetworkX ensures OPTWlogVOPT|OPT_W| \leq \log |V| \cdot OPT , and our algorithm guarantees OPTB2OPT|OPT_B| \leq 2 \cdot OPT , where OPTOPT is the optimal dominating set size (unknown in practice). Thus, the metric logVOPTBOPTW\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} provides insight into how close our solution is to the theoretical 2-approximation bound. A value near 2 indicates that our algorithm is performing near-optimally relative to the baseline.

Results and Analysis

The experimental results for a subset of the DIMACS instances are listed below. Each entry includes the dominating set size and runtime (in milliseconds) for our algorithm ( OPTB|OPT_B| ) and the NetworkX baseline ( OPTW|OPT_W| ), along with the computed approximation quality metric logVOPTBOPTW\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} .

  • p_hat500-1.clq: OPTB=10|OPT_B| = 10 (259.872 ms), OPTW=14|OPT_W| = 14 (37.267 ms), logVOPTBOPTW=4.439006\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} = 4.439006
  • p_hat500-2.clq: OPTB=5|OPT_B| = 5 (535.828 ms), OPTW=7|OPT_W| = 7 (31.832 ms), logVOPTBOPTW=4.439006\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} = 4.439006
  • p_hat500-3.clq: OPTB=3|OPT_B| = 3 (781.632 ms), OPTW=3|OPT_W| = 3 (19.985 ms), logVOPTBOPTW=6.214608\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} = 6.214608
  • p_hat700-1.clq: OPTB=10|OPT_B| = 10 (451.447 ms), OPTW=22|OPT_W| = 22 (70.044 ms), logVOPTBOPTW=2.977764\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} = 2.977764
  • p_hat700-2.clq: OPTB=6|OPT_B| = 6 (1311.585 ms), OPTW=8|OPT_W| = 8 (69.925 ms), logVOPTBOPTW=4.913310\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} = 4.913310
  • p_hat700-3.clq: OPTB=3|OPT_B| = 3 (1495.283 ms), OPTW=3|OPT_W| = 3 (72.238 ms), logVOPTBOPTW=6.551080\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} = 6.551080
  • san1000.clq: OPTB=4|OPT_B| = 4 (1959.679 ms), OPTW=40|OPT_W| = 40 (630.204 ms), logVOPTBOPTW=0.690776\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} = 0.690776
  • san200_0.7_1.clq: OPTB=3|OPT_B| = 3 (93.572 ms), OPTW=4|OPT_W| = 4 (5.196 ms), logVOPTBOPTW=3.973738\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} = 3.973738
  • san200_0.7_2.clq: OPTB=3|OPT_B| = 3 (103.698 ms), OPTW=5|OPT_W| = 5 (6.463 ms), logVOPTBOPTW=3.178990\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} = 3.178990
  • san200_0.9_1.clq: OPTB=2|OPT_B| = 2 (115.282 ms), OPTW=2|OPT_W| = 2 (0.000 ms), logVOPTBOPTW=5.298317\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} = 5.298317
  • san200_0.9_2.clq: OPTB=2|OPT_B| = 2 (120.091 ms), OPTW=2|OPT_W| = 2 (5.012 ms), logVOPTBOPTW=5.298317\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} = 5.298317
  • san200_0.9_3.clq: OPTB=2|OPT_B| = 2 (110.157 ms), OPTW=3|OPT_W| = 3 (0.000 ms), logVOPTBOPTW=3.532212\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} = 3.532212
  • san400_0.5_1.clq: OPTB=4|OPT_B| = 4 (243.552 ms), OPTW=24|OPT_W| = 24 (45.267 ms), logVOPTBOPTW=0.998577\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} = 0.998577
  • san400_0.7_1.clq: OPTB=3|OPT_B| = 3 (419.706 ms), OPTW=6|OPT_W| = 6 (20.579 ms), logVOPTBOPTW=2.995732\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} = 2.995732
  • san400_0.7_2.clq: OPTB=3|OPT_B| = 3 (405.550 ms), OPTW=6|OPT_W| = 6 (24.712 ms), logVOPTBOPTW=2.995732\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} = 2.995732
  • san400_0.7_3.clq: OPTB=3|OPT_B| = 3 (452.306 ms), OPTW=9|OPT_W| = 9 (33.302 ms), logVOPTBOPTW=1.997155\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} = 1.997155
  • san400_0.9_1.clq: OPTB=2|OPT_B| = 2 (453.124 ms), OPTW=3|OPT_W| = 3 (20.981 ms), logVOPTBOPTW=3.994310\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} = 3.994310
  • sanr200_0.7.clq: OPTB=3|OPT_B| = 3 (96.323 ms), OPTW=4|OPT_W| = 4 (7.047 ms), logVOPTBOPTW=3.973738\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} = 3.973738
  • sanr200_0.9.clq: OPTB=2|OPT_B| = 2 (116.587 ms), OPTW=2|OPT_W| = 2 (2.892 ms), logVOPTBOPTW=5.298317\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} = 5.298317
  • sanr400_0.5.clq: OPTB=6|OPT_B| = 6 (340.535 ms), OPTW=7|OPT_W| = 7 (20.473 ms), logVOPTBOPTW=5.135541\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} = 5.135541
  • sanr400_0.7.clq: OPTB=3|OPT_B| = 3 (490.877 ms), OPTW=5|OPT_W| = 5 (22.703 ms), logVOPTBOPTW=3.594879\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} = 3.594879

Our analysis of the results yields the following insights:

  • Runtime Efficiency: Our algorithm, implemented in Baldor, exhibits competitive runtime performance compared to NetworkX, particularly on larger instances like san1000.clq. However, NetworkX is generally faster on smaller graphs (e.g., san200_0.9_1.clq with a runtime of 0.000 ms), likely due to its simpler heuristic approach. In contrast, our algorithm’s runtime increases with graph size (e.g., 1959.679 ms for san1000.clq), reflecting the trade-off for achieving a better approximation guarantee. This suggests that while our algorithm is more computationally intensive, it scales reasonably well for the improved solution quality it provides.

  • Approximation Quality: The approximation quality metric logVOPTBOPTW\log |V| \cdot \frac{|OPT_B|}{|OPT_W|} frequently approaches the theoretical 2-approximation bound, with values such as 1.997155 for san400_0.7_3.clq and 2.977764 for p_hat700-1.clq. In cases like san1000.clq (0.690776), our algorithm significantly outperforms NetworkX, producing a dominating set of size 4 compared to NetworkX’s 40. However, for instances where OPTB=OPTW|OPT_B| = |OPT_W| (e.g., p_hat500-3.clq), the metric exceeds 2 due to the logarithmic factor, indicating that both algorithms may be far from the true optimum. Overall, our algorithm consistently achieves solutions closer to the theoretical optimum, validating its 2-approximation guarantee.

Discussion and Implications

The results highlight a favorable trade-off between solution quality and computational efficiency for our algorithm. On instances where approximation accuracy is critical, such as san1000.clq and san400_0.5_1.clq, our algorithm produces significantly smaller dominating sets than NetworkX, demonstrating its practical effectiveness. However, the increased runtime on larger graphs suggests opportunities for optimization, particularly in reducing redundant computations or leveraging parallelization.

These findings position our algorithm as a strong candidate for applications requiring high-quality approximations, such as network design, facility location, and clustering problems, where a 2-approximation guarantee can lead to substantial cost savings. For scenarios prioritizing speed over solution quality, the NetworkX baseline may be preferable due to its faster execution.

Future Work

Future research will focus on optimizing the runtime performance of our algorithm without compromising its approximation guarantees. Potential directions include:

  • Implementing heuristic-based pruning techniques to reduce the search space.
  • Exploring parallel and distributed computing to handle larger graphs more efficiently.
  • Extending the algorithm to handle weighted dominating set problems, broadening its applicability.

Additionally, we plan to evaluate our algorithm on real-world graphs from domains such as social networks and biological networks, where structural properties may further highlight the strengths of our approach.


References

  • [Johnson1996] Johnson, D. S., & Trick, M. A. (1996). Cliques, Coloring, and Satisfiability: Second DIMACS Implementation Challenge. DIMACS Series in Discrete Mathematics and Theoretical Computer Science.
  • [Vega25] Vega, D. (2025). Baldor: Approximate Minimum Dominating Set Solver (v0.1.3). Software Library.
  • [Vazirani2001] Vazirani, V. V. (2001). Approximation Algorithms. Springer.

Impact of This Result


Conclusion

These algorithms advance the field of approximation algorithms by balancing efficiency and solution quality, offering both theoretical depth and practical relevance. Their argument implies that P=NPP = NP would have far-reaching practical applications, particularly in artificial intelligence, medicine, and industrial sectors. This work is available as a PDF document on Preprints at the following link: https://www.preprints.org/manuscript/202504.0522/v1.