diff --git a/tests/framework/microvm.py b/tests/framework/microvm.py index d3a5fdceaf9..45375cd9c61 100644 --- a/tests/framework/microvm.py +++ b/tests/framework/microvm.py @@ -390,7 +390,7 @@ def _validate_api_response_times(self): "Got API call duration log entry before request entry" ) - if current_call.url != "/snapshot/create": + if current_call.url not in ["/snapshot/create", "/snapshot/load"]: exec_time = float(match.group("execution_time")) / 1000.0 assert ( diff --git a/tools/devtool b/tools/devtool index 44f10657c06..57637a553cc 100755 --- a/tools/devtool +++ b/tools/devtool @@ -743,9 +743,9 @@ cmd_test() { # It seems that even if the tests using huge pages run sequentially on ag=1 agents, right-sizing the huge pages # pool to the total number of huge pages used across all tests results in spurious failures with pool depletion # anyway (something else on the host seems to be stealing our huge pages, and we cannot "ear mark" them for - # Firecracker processes). Thus, just allocate 8GB of them and call it a day. + # Firecracker processes). Thus, just allocate 48GB of them and call it a day. say "Setting up huge pages pool" - num_hugetlbfs_pages=4096 + num_hugetlbfs_pages=24552 huge_pages_old=$(cat /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages) huge_pages_new=$(echo $num_hugetlbfs_pages |sudo tee /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages)