mean_test_time is the total time when a request is sent by a client until the client gets response ;
mean_time_to_first_byte is the time when a request is send until the first byte is received. including host resolving and server connecting time.
my question is:
1. are my above understandings correct?
2. mean_time_to_first_byte equals wait latency?
3. is there a method to get time_to_last_byte for httpplugin?
4. how to get mean throughput(the amount of data that clients receive from the server)