Hey Everyone,
I need to show a chart that displays between 0 & 120 Seconds (including milliseconds) for Server Loading times tested daily and, then average out the worst servers based upon the daily values entered.
In example;
06/15/2017 - 4.02 Seconds
06/16/2017 - 22.92 Seconds
06/17/2017 - 11.03 Seconds
06/18/2017 - 5.90 Seconds
06/19/2017 - 4.62 Seconds
Average Server Loading Time - 9.69 Seconds
I've tried it with time. but, it makes the graph & charts go crazy. It also seems to try and switch the numbers to an actual time as in 01/01/1901 00:10 AM which, also messes up the charts.
I then tried it as currency or a value(number) and even generic. But, it's always incorrectly displaying in charts & graphs.
What would be the best way to accomplish this?
I need to show a chart that displays between 0 & 120 Seconds (including milliseconds) for Server Loading times tested daily and, then average out the worst servers based upon the daily values entered.
In example;
06/15/2017 - 4.02 Seconds
06/16/2017 - 22.92 Seconds
06/17/2017 - 11.03 Seconds
06/18/2017 - 5.90 Seconds
06/19/2017 - 4.62 Seconds
Average Server Loading Time - 9.69 Seconds
I've tried it with time. but, it makes the graph & charts go crazy. It also seems to try and switch the numbers to an actual time as in 01/01/1901 00:10 AM which, also messes up the charts.
I then tried it as currency or a value(number) and even generic. But, it's always incorrectly displaying in charts & graphs.
What would be the best way to accomplish this?