|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH] Introduce an s3 test
On May 2, 2013, at 11:06 AM, Ian Jackson <Ian.Jackson@xxxxxxxxxxxxx>
wrote:
> Ben Guthro writes ("Re: [PATCH] Introduce an s3 test"):
>> On 05/01/2013 06:56 AM, Ian Jackson wrote:
>>>> +# check log for resume message
>>>> +poll_loop(4*$timeout, 2, 's3-confirm-resumed',
>>>> + target_cmd_output($ho,"xl dmesg | grep 'ACPI S' | tail -1 | " .
>>>> + "grep -n 'Finishing wakeup from S3 state'"));
>>>
>>> Why does this need a poll loop ? Surely after the machine comes out
>>> of suspend it should be up right away ?
>>
>> This is a bit of a "first pass" in a test environment I've never used
>> before. I modeled this after other tests I found in the same dir. If
>> this is inappropriate, then I suspect you are correct.
>
> Maybe you should be using guest_check_up ?
I'll own up to the fact that I wasn't really able to test the infrastructure
portions of this script.
I was unsuccessful in getting them to run, even using the "standalone" branch.
It would really help if someone who has access to the test infrastructure could
take my script as a starting point, and adapt it to whatever is necessary for
that test environment.
>
>> I put it in the loop for the case of networking taking some time to come
>> back online, so if the ssh command failed it would be retried.
>
> How long is it supposed to take to come back online ? "4*$timeout"
> seems (a) a bit arbitrary (b) rather long with your existing value of
> $timeout.
For all devices to come back online, it can sometimes take up to 20s.
This value was arbitrary, but chosen with the RTC variance + devices coming on
line.
This should probably be a tunable value.
>
>> Additionally, I have found that the RTC wakeup mechanism is not very
>> accurate in its timing.
>
> How unfortunate.
Indeed. We frequently see sleeping machines for 1m sometimes results in
sometimes results in machines waking up 30s later - others 3m later.
>
>>> I'm not sure I follow this. Wouldn't messed up timer queues cause
>>> other trouble in the guest ?
>>
>> Yes, but it has been a common point of failure / problems after S3. I
>> put this here as a placeholder to verify that everything is still as it
>> should be.
>
> Err, OK.
>
I see automated testing as a resource to be able to confirm that problems that
occurred in the past do not re-emerge from new development, rather than
strictly functional testing.
If you disagree with this, feel free to remove it. I don't feel strongly about
this particular point.
>>>> +# - Check for kernel Oops
>>>> +# - Check for Xen WARN
>>>
>>> These are a good idea but should perhaps be a separate test step.
>>
>> Wouldn't you want a warning/oops that was provoked by S3 to be
>> associated with that test?
>
> Hrm. Well in principle this is surely true of any test.
>
> Can we make warnings/oopses fatal ?
>
That seems like it would be prudent, if possible.
As I mentioned above, I had difficulty configuring this test environment, so it
may be trivial, and I am just not familiar enough with this environment.
Ben
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |