What does scheduler/aux mean in msacc?

Fernando Benavides elbrujohalcon@REDACTED
Tue Jul 7 13:52:50 CEST 2020


Thanks Jesper,

I understand all that but I already run the tests in more than 32 different
AWS machines… in groups of 2 OTP22 / 2 OTP21 machines each time… and the
results are impressively consistent. Every single time I get the results I
posted before (different %s of course, but not significantly different and
it's always the same difference ratio between the different OTP versions).
I'm really really convinced that this is not a fluke due to virtualization.
It's something that our system is doing that got worse in OTP22, for sure…
but I don't know where to look.

Cheers!

On Tue, Jul 7, 2020 at 12:40 PM Jesper Louis Andersen <
jesper.louis.andersen@REDACTED> wrote:

> You should run on the same machines, and preferably at the same time (back
> to back). When you are running under a hypervisor, you are cooperating with
> other users of the hardware, and they can wreak serious havoc on your
> benchmarks. I've seen 20% fluctuations in performance on AWS hardware due
> to this. The usual problems stems from how the CPUs handle caches and if
> there are other users who move a lot of data around, you will see a lower
> performance due to this.
>
> My advice will be to reconstruct the problem on local hardware. If you
> cannot do this, then chances are you are looking at some AWS artifact
> rather than a change in OTP behavior.
>
>
>
> On Mon, Jul 6, 2020 at 5:14 PM Fernando Benavides <elbrujohalcon@REDACTED>
> wrote:
>
>> Yeah… the curious thing is that both machines are the same kind of AWS
>> virtual machines… And I had 2 running OTP22 and 2 running OTP21… and both
>> OTP22 machines showed xen_hypercall_event_channel_op while both OTP21
>> didn't (actually… they did, but with a very smaller % of overhead). And,
>> BTW, the same happened with smp_call_function_many (more overhead
>> associated with that function in OTP22 than OTP21).
>>
>> I'll keep digging, but if anybody has any idea of what to look for, I
>> would appreciate it a lot.
>>
>> Thanks
>>
>> On Mon, Jul 6, 2020 at 5:05 PM Max Lapshin <max.lapshin@REDACTED> wrote:
>>
>>> First post from the internet about xen_hypercall_event_channel_op:
>>> https://www.scylladb.com/2018/06/21/impact-virtualization-database/
>>>
>>> TL;DR: metal server better than virtual
>>>
>>> On Mon, Jul 6, 2020 at 5:26 PM Fernando Benavides <
>>> elbrujohalcon@REDACTED> wrote:
>>>
>>>> I'll add another data point…
>>>> As Lukas suggested, I used *perf* on the running systems and this is
>>>> what I've got…
>>>>
>>>> For OTP22:
>>>>      2.33%  8_scheduler      beam.smp             [.] process_main
>>>>
>>>> *     …the rest of the schedulers running process_main…*     1.57%
>>>>  6_scheduler      beam.smp             [.] process_main
>>>>      0.34%  6_scheduler      [kernel.kallsyms]    [k]
>>>> xen_hypercall_event_channel_op
>>>>      0.32%  7_scheduler      [kernel.kallsyms]    [k]
>>>> xen_hypercall_event_channel_op
>>>>      0.25%  14_scheduler     beam.smp             [.] copy_struct_x
>>>>      0.25%  2_scheduler      beam.smp             [.] copy_struct_x
>>>>      0.25%  9_scheduler      [kernel.kallsyms]    [k]
>>>> xen_hypercall_event_channel_op
>>>>      0.23%  11_scheduler     [kernel.kallsyms]    [k]
>>>> xen_hypercall_event_channel_op
>>>>      0.22%  11_scheduler     beam.smp             [.] copy_shallow_x
>>>>      0.22%  9_scheduler      libc-2.17.so         [.]
>>>> __memcpy_ssse3_back
>>>>      0.20%  13_scheduler     [kernel.kallsyms]    [k] __lock_text_start
>>>>      0.20%  3_scheduler      [kernel.kallsyms]    [k]
>>>> xen_hypercall_event_channel_op
>>>>      0.19%  10_scheduler     beam.smp             [.] erts_cmp_compound
>>>>
>>>> For OTP21:
>>>>      2.45%  4_scheduler      beam.smp             [.] process_main
>>>>
>>>> *     …the rest of the schedulers running process_main…*     1.91%
>>>>  5_scheduler      beam.smp             [.] process_main
>>>>      0.45%  15_scheduler     [kernel.kallsyms]    [k] __lock_text_start
>>>>      0.27%  10_scheduler     [kernel.kallsyms]    [k] __lock_text_start
>>>>      0.27%  3_scheduler      [kernel.kallsyms]    [k] __lock_text_start
>>>>      0.27%  7_scheduler      [kernel.kallsyms]    [k] __lock_text_start
>>>>      0.25%  1_scheduler      beam.smp             [.] copy_struct_x
>>>>      0.25%  2_scheduler      [kernel.kallsyms]    [k] __lock_text_start
>>>>      0.23%  2_scheduler      beam.smp             [.] copy_struct_x
>>>>      0.23%  7_scheduler      beam.smp             [.] copy_struct_x
>>>>      0.20%  4_scheduler      beam.smp             [.] copy_struct_x
>>>>      0.20%  6_scheduler      [kernel.kallsyms]    [k] __lock_text_start
>>>>      0.18%  14_scheduler     beam.smp             [.] copy_struct_x
>>>>
>>>> I'm not sure what any of those functions are, or why
>>>> xen_hypercall_event_channel_op is so much popular in OTP22, tho… ��
>>>>
>>>> On Mon, Jul 6, 2020 at 3:42 PM Max Lapshin <max.lapshin@REDACTED>
>>>> wrote:
>>>>
>>>>> In a nutshell there is a non-reproducible fluctuation of measurement:
>>>>> one subsystem takes 10% instead 8%, another 21% instead 23%.
>>>>>
>>>>> Totally there seems to be no change.
>>>>>
>>>>> It is extremely hard to make reliable measurement that helps to find
>>>>> 8% to 10% change.
>>>>>
>>>>
>>>>
>>>> --
>>>>
>>>> <https://about.me/elbrujohalcon?promo=email_sig&utm_source=product&utm_medium=email_sig&utm_campaign=gmail_api&utm_content=thumb>
>>>> Brujo Benavides
>>>> about.me/elbrujohalcon
>>>> <https://about.me/elbrujohalcon?promo=email_sig&utm_source=product&utm_medium=email_sig&utm_campaign=gmail_api&utm_content=thumb>
>>>>
>>>
>>
>> --
>>
>> <https://about.me/elbrujohalcon?promo=email_sig&utm_source=product&utm_medium=email_sig&utm_campaign=gmail_api&utm_content=thumb>
>> Brujo Benavides
>> about.me/elbrujohalcon
>> <https://about.me/elbrujohalcon?promo=email_sig&utm_source=product&utm_medium=email_sig&utm_campaign=gmail_api&utm_content=thumb>
>>
>
>
> --
> J.
>


-- 
<https://about.me/elbrujohalcon?promo=email_sig&utm_source=product&utm_medium=email_sig&utm_campaign=gmail_api&utm_content=thumb>
Brujo Benavides
about.me/elbrujohalcon
<https://about.me/elbrujohalcon?promo=email_sig&utm_source=product&utm_medium=email_sig&utm_campaign=gmail_api&utm_content=thumb>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://erlang.org/pipermail/erlang-questions/attachments/20200707/9aa87bc6/attachment.htm>


More information about the erlang-questions mailing list