Typical notataion about the algorithm complexity and some thoughts
when consider a specific optimization, it might makes sense if it decrease the upper bound, otherwise, the optimization in the constant item maybe meaningless. When considering the optimization for a system, it might makes sense if you consider it from the complexity aspects.
When we use big-Θ notation, we’re saying that we have an asymptotically tight bound on the running time. “Asymptotically” because it matters for only large values of nnn. “Tight bound” because we’ve nailed the running time to within a constant factor above and below.
big-omega, asymptotic upper bound, It would be convenient to have a form of asymptotic notation that means “the running time grows at most this much, but it could grow more slowly.” We use “big-O” notation for just such occasions.
For example, although the worst-case running time of binary search is big theta (log_2n), it would be incorrect to say that the binary search runs in big theta (log_2n), sometimes, we could find suitable value at first guess, the running time will be less then the (log_2n). Therefore, we could use the big O(log_2n) to represent this scenario.
We may much care about the upper bound for most algorithms.
We use big-Ω notation for asymptotic lower bounds, since it bounds the growth of the running time from below for large enough input sizes.
The consideration about the notation explains why it is important to consider the metrics for research work. you always need to improve particular metric in most of the cases. How time can be saved? use better algoeithm, use parallel resources or reduce unnecessary things. It is slways good to consider what metric are you using. I may use the workflow running time previously, but I found that it might be better to consider the overhead, since we could not decrease the time spent on the simulation, and we do not instrument into the simulation actually.
There are different angle when there are components running concurently, the bottleneck might be the important one. consider the simulation and analytics case. if the simulation its self is the bottleneck for the whole workflow, it might not much useful to decrease the analytics time or other related overhead. Another issue is the globle optimizaiton or the local optimization, namely if you need to anticipate the future or just current information are ok.
I may get lost sometimes, things the project I do may not have a good metric or performance improvement, some times I even can find a such metric. Maybe consider from the application’s perspective is really important, try to do sth that is good for particular application. This is an important knife to provides a new solutions.