| #!/usr/bin/env python3.4 |
| # |
| # Copyright 2017 - The Android Open Source Project |
| # |
| # Licensed under the Apache License, Version 2.0 (the "License"); |
| # you may not use this file except in compliance with the License. |
| # You may obtain a copy of the License at |
| # |
| # http://www.apache.org/licenses/LICENSE-2.0 |
| # |
| # Unless required by applicable law or agreed to in writing, software |
| # distributed under the License is distributed on an "AS IS" BASIS, |
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| # See the License for the specific language governing permissions and |
| # limitations under the License. |
| |
| import collections |
| import json |
| import logging |
| import os |
| import WifiRvrTest |
| from acts import asserts |
| from acts import base_test |
| from acts import utils |
| from acts.test_decorators import test_tracker_info |
| from acts.test_utils.wifi import wifi_test_utils as wutils |
| from acts.test_utils.wifi import wifi_retail_ap as retail_ap |
| |
| |
| class WifiSensitivityTest(WifiRvrTest.WifiRvrTest): |
| """Class to test WiFi sensitivity tests. |
| |
| This class implements measures WiFi sensitivity per rate. It heavily |
| leverages the WifiRvrTest class and introduced minor differences to set |
| specific rates and the access point, and implements a different pass/fail |
| check. For an example config file to run this test class see |
| example_connectivity_performance_ap_sta.json. |
| """ |
| |
| VALID_TEST_CONFIGS = { |
| 1: ["legacy", "VHT20"], |
| 2: ["legacy", "VHT20"], |
| 6: ["legacy", "VHT20"], |
| 10: ["legacy", "VHT20"], |
| 11: ["legacy", "VHT20"], |
| 36: ["legacy", "VHT20"], |
| 40: ["legacy", "VHT20"], |
| 44: ["legacy", "VHT20"], |
| 48: ["legacy", "VHT20"], |
| 149: ["legacy", "VHT20"], |
| 153: ["legacy", "VHT20"], |
| 157: ["legacy", "VHT20"], |
| 161: ["legacy", "VHT20"] |
| } |
| VALID_RATES = { |
| "legacy_2GHz": [[54, 1], [48, 1], [36, 1], [24, 1], [18, 1], [12, 1], |
| [11, 1], [9, 1], [6, 1], [5.5, 1], [2, 1], [1, 1]], |
| "legacy_5GHz": [[54, 1], [48, 1], [36, 1], [24, 1], [18, 1], [12, 1], |
| [9, 1], [6, 1]], |
| "HT": [[8, 1], [7, 1], [6, 1], [5, 1], [4, 1], [3, 1], [2, 1], [1, 1], |
| [0, 1], [15, 2], [14, 2], [13, 2], [12, 2], [11, 2], [10, 2], |
| [9, 2], [8, 2]], |
| "VHT": [[9, 1], [8, 1], [7, 1], [6, 1], [5, 1], [4, 1], [3, 1], [2, 1], |
| [1, 1], [0, 1], [9, 2], [8, 2], [7, 2], [6, 2], [5, 2], [4, 2], |
| [3, 2], [2, 2], [1, 2], [0, 2]] |
| } |
| |
| def setup_class(self): |
| """Initializes common test hardware and parameters. |
| |
| This function initializes hardwares and compiles parameters that are |
| common to all tests in this class. |
| """ |
| self.client_dut = self.android_devices[-1] |
| req_params = [ |
| "RetailAccessPoints", "sensitivity_test_params", "testbed_params" |
| ] |
| opt_params = ["main_network", "golden_files_list"] |
| self.unpack_userparams(req_params, opt_params) |
| self.testclass_params = self.sensitivity_test_params |
| self.num_atten = self.attenuators[0].instrument.num_atten |
| self.iperf_server = self.iperf_servers[0] |
| self.access_points = retail_ap.create(self.RetailAccessPoints) |
| self.access_point = self.access_points[0] |
| self.log.info("Access Point Configuration: {}".format( |
| self.access_point.ap_settings)) |
| self.log_path = os.path.join(logging.log_path, "results") |
| utils.create_dir(self.log_path) |
| if not hasattr(self, "golden_files_list"): |
| self.golden_files_list = [ |
| os.path.join(self.testbed_params["golden_results_path"], |
| file) for file in os.listdir( |
| self.testbed_params["golden_results_path"]) |
| ] |
| self.testclass_results = [] |
| |
| # Turn WiFi ON |
| for dev in self.android_devices: |
| wutils.wifi_toggle_state(dev, True) |
| |
| def pass_fail_check(self, rvr_result): |
| """Checks sensitivity against golden results and decides on pass/fail. |
| |
| Args: |
| rvr_result: dict containing attenuation, throughput and other meta |
| data |
| """ |
| try: |
| golden_path = next(file_name |
| for file_name in self.golden_files_list |
| if "sensitivity_targets" in file_name) |
| with open(golden_path, 'r') as golden_file: |
| golden_results = json.load(golden_file) |
| except: |
| asserts.fail("Test failed: Golden file not found") |
| |
| golden_sensitivity = golden_results[self.current_test_name][ |
| "sensitivity"] |
| result_string = "Througput = {}, Sensitivity = {}. Target Sensitivity = {}".format( |
| rvr_result["peak_throughput"], rvr_result["sensitivity"], |
| golden_sensitivity) |
| if rvr_result["sensitivity"] - golden_sensitivity > self.testclass_params["sensitivity_tolerance"]: |
| asserts.fail("Test Failed. {}".format(result_string)) |
| else: |
| asserts.explicit_pass("Test Passed. {}".format(result_string)) |
| |
| def process_testclass_results(self): |
| """Saves and plots test results from all executed test cases.""" |
| testclass_results_dict = collections.OrderedDict() |
| for result in self.testclass_results: |
| testclass_results_dict[result["test_name"]] = { |
| "peak_throughput": result["peak_throughput"], |
| "range": result["range"], |
| "sensitivity": result["sensitivity"] |
| } |
| results_file_path = os.path.join(self.log_path, 'results.json') |
| with open(results_file_path, 'w') as results_file: |
| json.dump(testclass_results_dict, results_file, indent=4) |
| super().process_testclass_results() |
| |
| def process_test_results(self, rvr_result): |
| """Post processes RvR results to compute sensitivity. |
| |
| Takes in the results of the RvR tests and computes the sensitivity of |
| the current rate by looking at the point at which throughput drops |
| below the percentage specified in the config file. The function then |
| calls on its parent class process_test_results to plot the result. |
| |
| Args: |
| rvr_result: dict containing attenuation, throughput and other meta |
| data |
| """ |
| rvr_result["peak_throughput"] = max(rvr_result["throughput_receive"]) |
| throughput_check = [ |
| throughput < rvr_result["peak_throughput"] * |
| (self.testclass_params["throughput_pct_at_sensitivity"] / 100) |
| for throughput in rvr_result["throughput_receive"] |
| ] |
| consistency_check = [ |
| idx for idx in range(len(throughput_check)) |
| if all(throughput_check[idx:]) |
| ] |
| rvr_result["atten_at_range"] = rvr_result["attenuation"][ |
| consistency_check[0] - 1] |
| rvr_result["range"] = rvr_result["fixed_attenuation"] + ( |
| rvr_result["atten_at_range"]) |
| rvr_result["sensitivity"] = self.testclass_params["ap_tx_power"] - ( |
| rvr_result["range"]) |
| super().process_test_results(rvr_result) |
| |
| def setup_ap(self, testcase_params): |
| """Sets up the access point in the configuration required by the test. |
| |
| Args: |
| testcase_params: dict containing AP and other test params |
| """ |
| band = self.access_point.band_lookup_by_channel( |
| testcase_params["channel"]) |
| if "2G" in band: |
| frequency = wutils.WifiEnums.channel_2G_to_freq[testcase_params[ |
| "channel"]] |
| else: |
| frequency = wutils.WifiEnums.channel_5G_to_freq[testcase_params[ |
| "channel"]] |
| if frequency in wutils.WifiEnums.DFS_5G_FREQUENCIES: |
| self.access_point.set_region(self.testbed_params["DFS_region"]) |
| else: |
| self.access_point.set_region(self.testbed_params["default_region"]) |
| self.access_point.set_channel(band, testcase_params["channel"]) |
| self.access_point.set_bandwidth(band, testcase_params["mode"]) |
| self.access_point.set_power(band, testcase_params["ap_tx_power"]) |
| self.access_point.set_rate( |
| band, testcase_params["mode"], testcase_params["num_streams"], |
| testcase_params["rate"], testcase_params["short_gi"]) |
| self.log.info("Access Point Configuration: {}".format( |
| self.access_point.ap_settings)) |
| |
| def get_start_atten(self): |
| """Gets the starting attenuation for this sensitivity test. |
| |
| The function gets the starting attenuation by checking whether a test |
| as the next higher MCS has been executed. If so it sets the starting |
| point a configurable number of dBs below the next MCS's sensitivity. |
| |
| Returns: |
| start_atten: starting attenuation for current test |
| """ |
| # Get the current and reference test config. The reference test is the |
| # one performed at the current MCS+1 |
| current_test_params = self.parse_test_params(self.current_test_name) |
| ref_test_params = current_test_params.copy() |
| if "legacy" in current_test_params["mode"] and current_test_params["rate"] < 54: |
| if current_test_params["channel"] <= 13: |
| ref_index = self.VALID_RATES["legacy_2GHz"].index( |
| [current_test_params["rate"], 1]) - 1 |
| ref_test_params["rate"] = self.VALID_RATES["legacy_2GHz"][ |
| ref_index][0] |
| else: |
| ref_index = self.VALID_RATES["legacy_5GHz"].index( |
| [current_test_params["rate"], 1]) - 1 |
| ref_test_params["rate"] = self.VALID_RATES["legacy_5GHz"][ |
| ref_index][0] |
| else: |
| ref_test_params["rate"] = ref_test_params["rate"] + 1 |
| |
| # Check if reference test has been run and set attenuation accordingly |
| previous_params = [ |
| self.parse_test_params(result["test_name"]) |
| for result in self.testclass_results |
| ] |
| try: |
| ref_index = previous_params.index(ref_test_params) |
| start_atten = self.testclass_results[ref_index]["atten_at_range"] - ( |
| self.testclass_params["adjacent_mcs_range_gap"]) |
| except: |
| print("Reference test not found. Starting from {} dB".format( |
| self.testclass_params["atten_start"])) |
| start_atten = self.testclass_params["atten_start"] |
| return start_atten |
| |
| def parse_test_params(self, test_name): |
| """Function that generates test params based on the test name.""" |
| test_name_params = test_name.split("_") |
| testcase_params = collections.OrderedDict() |
| testcase_params["channel"] = int(test_name_params[2][2:]) |
| testcase_params["mode"] = test_name_params[3] |
| if "legacy" in testcase_params["mode"].lower(): |
| testcase_params["rate"] = float( |
| str(test_name_params[4]).replace("p", ".")) |
| else: |
| testcase_params["rate"] = int(test_name_params[4][3:]) |
| testcase_params["num_streams"] = int(test_name_params[5][3:]) |
| testcase_params["short_gi"] = 0 |
| if self.testclass_params["traffic_type"] == "UDP": |
| testcase_params["iperf_args"] = '-i 1 -t {} -J -u -b {} -R'.format( |
| self.testclass_params["iperf_duration"], |
| self.testclass_params["UDP_rates"][testcase_params["mode"]]) |
| else: |
| testcase_params["iperf_args"] = '-i 1 -t {} -J -R'.format( |
| self.testclass_params["iperf_duration"]) |
| testcase_params["use_client_output"] = True |
| return testcase_params |
| |
| def _test_sensitivity(self): |
| """ Function that gets called for each test case |
| |
| The function gets called in each rvr test case. The function customizes |
| the rvr test based on the test name of the test that called it |
| """ |
| # Compile test parameters from config and test name |
| testcase_params = self.parse_test_params(self.current_test_name) |
| testcase_params.update(self.testclass_params) |
| testcase_params["atten_start"] = self.get_start_atten() |
| |
| # Prepare devices and run test |
| self.setup_rvr_test(testcase_params) |
| rvr_result = self.run_rvr_test(testcase_params) |
| |
| # Post-process results |
| self.process_test_results(rvr_result) |
| self.testclass_results.append(rvr_result) |
| self.pass_fail_check(rvr_result) |
| # Add results to testclass_results |
| |
| def generate_test_cases(self, channels): |
| """Function that auto-generates test cases for a test class.""" |
| testcase_wrapper = self._test_sensitivity |
| for channel in channels: |
| for mode in self.VALID_TEST_CONFIGS[channel]: |
| if "VHT" in mode: |
| rates = self.VALID_RATES["VHT"] |
| elif "HT" in mode: |
| rates = self.VALID_RATES["HT"] |
| elif "legacy" in mode and channel < 14: |
| rates = self.VALID_RATES["legacy_2GHz"] |
| elif "legacy" in mode and channel > 14: |
| rates = self.VALID_RATES["legacy_5GHz"] |
| else: |
| raise ValueError("Invalid test mode.") |
| for rate in rates: |
| if "legacy" in mode: |
| testcase_name = "test_sensitivity_ch{}_{}_{}_nss{}".format( |
| channel, mode, |
| str(rate[0]).replace(".", "p"), rate[1]) |
| else: |
| testcase_name = "test_sensitivity_ch{}_{}_mcs{}_nss{}".format( |
| channel, mode, rate[0], rate[1]) |
| setattr(self, testcase_name, testcase_wrapper) |
| self.tests.append(testcase_name) |
| |
| |
| class WifiSensitivity_AllChannels_Test(WifiSensitivityTest): |
| def __init__(self, controllers): |
| base_test.BaseTestClass.__init__(self, controllers) |
| self.generate_test_cases( |
| [1, 2, 6, 10, 11, 36, 40, 44, 48, 149, 153, 157, 161]) |
| |
| |
| class WifiSensitivity_2GHz_Test(WifiSensitivityTest): |
| def __init__(self, controllers): |
| base_test.BaseTestClass.__init__(self, controllers) |
| self.generate_test_cases([1, 2, 6, 10, 11]) |
| |
| |
| class WifiSensitivity_UNII1_Test(WifiSensitivityTest): |
| def __init__(self, controllers): |
| base_test.BaseTestClass.__init__(self, controllers) |
| self.generate_test_cases([36, 40, 44, 48]) |
| |
| |
| class WifiSensitivity_UNII3_Test(WifiSensitivityTest): |
| def __init__(self, controllers): |
| base_test.BaseTestClass.__init__(self, controllers) |
| self.generate_test_cases([149, 153, 157, 161]) |
| |
| |
| class WifiSensitivity_ch1_Test(WifiSensitivityTest): |
| def __init__(self, controllers): |
| base_test.BaseTestClass.__init__(self, controllers) |
| self.generate_test_cases([1]) |
| |
| |
| class WifiSensitivity_ch2_Test(WifiSensitivityTest): |
| def __init__(self, controllers): |
| base_test.BaseTestClass.__init__(self, controllers) |
| self.generate_test_cases([2]) |
| |
| |
| class WifiSensitivity_ch6_Test(WifiSensitivityTest): |
| def __init__(self, controllers): |
| base_test.BaseTestClass.__init__(self, controllers) |
| self.generate_test_cases([6]) |
| |
| |
| class WifiSensitivity_ch10_Test(WifiSensitivityTest): |
| def __init__(self, controllers): |
| base_test.BaseTestClass.__init__(self, controllers) |
| self.generate_test_cases([10]) |
| |
| |
| class WifiSensitivity_ch11_Test(WifiSensitivityTest): |
| def __init__(self, controllers): |
| base_test.BaseTestClass.__init__(self, controllers) |
| self.generate_test_cases([11]) |
| |
| |
| class WifiSensitivity_ch36_Test(WifiSensitivityTest): |
| def __init__(self, controllers): |
| base_test.BaseTestClass.__init__(self, controllers) |
| self.generate_test_cases([36]) |
| |
| |
| class WifiSensitivity_ch40_Test(WifiSensitivityTest): |
| def __init__(self, controllers): |
| base_test.BaseTestClass.__init__(self, controllers) |
| self.generate_test_cases([40]) |
| |
| |
| class WifiSensitivity_ch44_Test(WifiSensitivityTest): |
| def __init__(self, controllers): |
| base_test.BaseTestClass.__init__(self, controllers) |
| self.generate_test_cases([44]) |
| |
| |
| class WifiSensitivity_ch48_Test(WifiSensitivityTest): |
| def __init__(self, controllers): |
| base_test.BaseTestClass.__init__(self, controllers) |
| self.generate_test_cases([48]) |
| |
| |
| class WifiSensitivity_ch149_Test(WifiSensitivityTest): |
| def __init__(self, controllers): |
| base_test.BaseTestClass.__init__(self, controllers) |
| self.generate_test_cases([149]) |
| |
| |
| class WifiSensitivity_ch153_Test(WifiSensitivityTest): |
| def __init__(self, controllers): |
| base_test.BaseTestClass.__init__(self, controllers) |
| self.generate_test_cases([153]) |
| |
| |
| class WifiSensitivity_ch157_Test(WifiSensitivityTest): |
| def __init__(self, controllers): |
| base_test.BaseTestClass.__init__(self, controllers) |
| self.generate_test_cases([157]) |
| |
| |
| class WifiSensitivity_ch161_Test(WifiSensitivityTest): |
| def __init__(self, controllers): |
| base_test.BaseTestClass.__init__(self, controllers) |
| self.generate_test_cases([161]) |