You have optimized an algorithm to improve its time complexity from O(n^2) to O(n log n). When you benchmark the optimized code on real-world data, you find that it's only slightly faster for small inputs and doesn't show significant improvement for larger datasets. What's the most likely explanation?