Note - Currently working on this blog
Client & Server Pattern
Master & Worker Pattern
Partition For Parallelism
Common Approach To Improve Throughput
Performance Optimization Concepts
Latency = time delay between starting an activity and when the results are available / detectable
Throughput = ratio of number of tasks completed in a unit of time.
Performance (perceived speed / responsiveness) = number of requests made and acknowledged in a unit of time.
Throughput and Performance are often confused! (sometimes they are the same)
Example:
Average Throughput = 10 tasks / sec
Average Latency = 1000ms (1sec / 10)
Performance = unknown
NOTE -
To improve performance = reduce latencies (between request and response)
To improve throughput = increase capacity (or reduce total latency)
Only Execute Mandatory Tasks
Increase CPU speed (scale up)
Optimize Algorithm
Exploit Parallelism (Scale Out)
Optimize Large Latencies
Best of all world - Do them all
Reduce Use Of XML
Use Case - Oracle Coherence In Compute Grid
No comments:
Post a Comment