Exam 2 Review
Reductions
We say that a problem reduces to a problem if there is a polynomial time reduction function such that for all , .
To prove a reduction, we need to show that the reduction function :
- runs in polynomial time
- .
Useful results from reductions
- is at least as hard as if .
- If we can solve in polynomial time, then we can solve in polynomial time.
- If we want to solve problem , and we already know an efficient algorithm for , then we can use the reduction to solve efficiently.
- If we want to show that is NP-hard, we can do this by showing that for some known NP-hard problem .
is the class of problems that can be solved in polynomial time. is the class of problems that can be verified in polynomial time.
We know that .
NP-complete problems
A problem is NP-complete if it is in and it is also NP-hard.
NP
A problem is in if
- there is a polynomial size certificate for the problem, and
- there is a polynomial time verifier for the problem that takes the certificate and checks whether it is a valid solution.
NP-hard
A problem is NP-hard if every instance of hard problem can be reduced to it in polynomial time.
List of known NP-hard problems:
- 3-SAT (or SAT):
- Statement: Given a boolean formula in CNF with at most 3 literals per clause, is there an assignment of truth values to the variables that makes the formula true?
- Independent Set:
- Statement: Given a graph and an integer , does contain a set of vertices such that no two vertices in the set are adjacent?
- Vertex Cover:
- Statement: Given a graph and an integer , does contain a set of vertices such that every edge in is incident to at least one vertex in the set?
- 3-coloring:
- Statement: Given a graph , can each vertex be assigned one of 3 colors such that no two adjacent vertices have the same color?
- Hamiltonian Cycle:
- Statement: Given a graph , does contain a cycle that visits every vertex exactly once?
- Hamiltonian Path:
- Statement: Given a graph , does contain a path that visits every vertex exactly once?
Approximation Algorithms
- Consider optimization problems whose decision problem variant is NP-hard. Unless P=NP, finding an optimal solution to these problems cannot be done in polynomial time.
- In approximation algorithms, we make a trade-o↵: we’re willing to accept sub-optimal solutions in exchange for polynomial runtime.
- The Approximation Ratio of our algorithm is the worst-case ratio of our solution to the optimal solution.
- For minimization problems, this ratio is , since our solution will be larger than OPT.
- For maximization problems, this ratio is , since our solution will be smaller than OPT.
- If given an algorithm, and you need to show it has some desired approximation ratio, there are a few approaches.
- In recitation, we saw Max-Subset Sum. We found upper bounds on the optimal solution and showed that the given algorithm would always give a solution with value at least half of the upper bound, giving our approximation ratio of 2.
- In lecture, you saw the Vertex Cover 2-approximation. Here, you would select any uncovered edge and add both u and v to the cover. We argued that at least one of u or v must be in the optimal cover, as the edge must be covered, so at every step we added at least one vertex from an optimal solution, and potentially one extra. So, the size of our cover could not be any larger than twice the optimal.
Randomized Algorithms
Sometimes, we can get better expected performance from an algorithm by introducing randomness.
We make the tradeoff guarantee runtime and solution quality from a deterministic algorithm, to expected runtime and quality from randomized algorithms.
We can make various bounds and tricks to calculate and amplify the probability of succeeding.
Chernoff Bound
Statement:
Requirements:
- is the sum of independent random variables
- You used the Chernoff bound to bound the probability of getting less than good partitions, since the probability of each partition being good is independent – the quality of one partition does not affect the quality of the next.
- Usage: If you have some probability that you want to bound, you must find , and find a value for such that . You can then plug in and into the Chernoff bound.
Markov’s Inequality
Statement:
Requirements:
- is a non-negative random variable
- No assumptions about independence
- Usage: If you have some probability that you want to bound, you must find , and find a value for such that . You can then plug in and into Markov’s inequality.
Union Bound
Statement:
- Conceptually, it’s saying that at least one event out of a collection will occur is no more than the sum of the probabilities of each event.
- Usage: To bound some bad event , we can use the union bound to sum up the probabilities of each of the bad events and use that to bound .
Probabilistic Boosting via Repeated Trials
- If we want to reduce the probability of some bad event to some value , we can run the algorithm repeatedly and make majority votes for the decision.
- Assume we run the algorithm times, and the probability of success is .
- The probability that all trials fail is at most .
- The majority vote of runs is wrong is the same as probability that more than trials fail.
- So, the probability is
- If we want this probability to be at most , we can just solve for in the inequality make it less than some . Then we solve for in the inequality .
Online Algorithms
- We make decisions on the fly, without knowing the future.
- The offline optimum is the optimal solution that knows the future.
- The competitive ratio of an online algorithm is the worst-case ratio of the cost of the online algorithm to the cost of the offline optimum. (when offline problem is NP-complete, an online algorithm for the problem is also an approximation algorithm)
- We do case by case analysis to show that the competitive ratio is at most some value. Just like approximation ratio proofs.