Re: [Gems-users] Abnormally high supervisor/user access ratio


Date: Thu, 8 Apr 2010 12:04:13 +0900
From: Ikhwan Lee <ikhwan@xxxxxxxxxxxxxxx>
Subject: Re: [Gems-users] Abnormally high supervisor/user access ratio
Thanks for you advice David,
 
Can you tell my which target machine you were using and what the daemon process was? I'm running SunOS abisko 5.10 on Serengeti, and I'm not seeing any high load processes on my top.

Regards,
Ikhwan
On Tue, Apr 6, 2010 at 11:27 PM, David Nellans <dnellans@xxxxxxxxxxx> wrote:
Assuming you are running this under Solaris...  in our experience Solaris has a nasty habit of
starting a daemon that will run at 100% cpu utilization for a long time in the background
occasionally.  We have also seen significantly skewed stats as a result of this similar to what
you're describing below.

To check, just fire up your checkpoint (no gems/opal required) kill your running app
and just watch top or the load monitor.  if that isn't zero when you're sitting idle - you're likely
having the same problem.



On Tue, 2010-04-06 at 10:10 -0400, Polina Dudnik wrote:
Ikhwan,

Unfortunately, we are not in control of Simics, which is what
determines the superviser/user accesses. So, your guess is as good as
ours.

Polina


On Fri, Apr 2, 2010 at 7:12 PM, Ikhwan Lee <ikhwan@xxxxxxxxxxxxxxx> wrote:
> I've seen similar issues reported in the past, but could not find a
> good answer. I would appreciate it if someone could explain the
> results below.
>
> I'm running Splash-2 programs on a 16-core RUBY+OPAL simulation
> setting. Some of the important Ruby parameters are as follows:
>
> protocol: MOESI_SMP_directory
> simics_version: Simics 3.0.31
> OPAL_RUBY_MULTIPLIER: 2
> L1_CACHE_ASSOC: 2                        // 2KB
> L1_CACHE_NUM_SETS_BITS: 4
> L2_CACHE_ASSOC: 4                       // 16KB
> L2_CACHE_NUM_SETS_BITS: 6
> g_NUM_PROCESSORS: 16
> g_NUM_L2_BANKS: 16
> g_NUM_MEMORIES: 16
> g_NUM_CHIPS: 16
> g_NETWORK_TOPOLOGY: FILE_SPECIFIED   // 4x4 mesh
> g_GARNET_NETWORK: true
> g_DETAIL_NETWORK: true
>
>
> When I grep SuperviserMode access types in the ruby stat files, almost
> all Splash-2 programs show excessively high superviser/user access
> ratio as you can see.
>
> BARNES/ruby.BARNES.16k.16p.16m.stats:
> L1D_cache_access_mode_type_SupervisorMode:   266621884    85.2083%
> BARNES/ruby.BARNES.16k.16p.16m.stats:
> L1I_cache_access_mode_type_SupervisorMode:   15695183    82.9426%
> BARNES/ruby.BARNES.16k.16p.16m.stats:
> L2_cache_access_mode_type_SupervisorMode:   145082595    83.6042%
> CHOLESKY/ruby.CHOLESKY.tk15.O.16p.16m.stats:
> L1D_cache_access_mode_type_SupervisorMode:   12393234    76.1279%
> CHOLESKY/ruby.CHOLESKY.tk15.O.16p.16m.stats:
> L1I_cache_access_mode_type_SupervisorMode:   10213059    88.0104%
> CHOLESKY/ruby.CHOLESKY.tk15.O.16p.16m.stats:
> L2_cache_access_mode_type_SupervisorMode:   5567772    82.6523%
> FFT/ruby.FFT.64k.16p.16m.stats:
> L1D_cache_access_mode_type_SupervisorMode:   4967929    87.4088%
> FFT/ruby.FFT.64k.16p.16m.stats:
> L1I_cache_access_mode_type_SupervisorMode:   353405    90.6429%
> FFT/ruby.FFT.64k.16p.16m.stats:
> L2_cache_access_mode_type_SupervisorMode:   2537530    92.726%
> FMM/ruby.FMM.16k.16p.16m.stats:
> L1D_cache_access_mode_type_SupervisorMode:   285541702    76.1547%
> FMM/ruby.FMM.16k.16p.16m.stats:
> L1I_cache_access_mode_type_SupervisorMode:   9056962    74.9888%
> FMM/ruby.FMM.16k.16p.16m.stats:
> L2_cache_access_mode_type_SupervisorMode:   109348107    93.9731%
> LUcon/ruby.FMM.16k.16p.16m.stats:
> L1D_cache_access_mode_type_SupervisorMode:   285541702    76.1547%
> LUcon/ruby.FMM.16k.16p.16m.stats:
> L1I_cache_access_mode_type_SupervisorMode:   9056962    74.9888%
> LUcon/ruby.FMM.16k.16p.16m.stats:
> L2_cache_access_mode_type_SupervisorMode:   109348107    93.9731%
> OCEAN_CONTIGUOUS/ruby.OCEAN_CONTIGUOUS.258.16p.16m.stats:
> L1D_cache_access_mode_type_SupervisorMode:   175987256    91.1572%
> OCEAN_CONTIGUOUS/ruby.OCEAN_CONTIGUOUS.258.16p.16m.stats:
> L1I_cache_access_mode_type_SupervisorMode:   12561588    85.7416%
> OCEAN_CONTIGUOUS/ruby.OCEAN_CONTIGUOUS.258.16p.16m.stats:
> L2_cache_access_mode_type_SupervisorMode:   91914021    89.7628%
> RADIOSITY/ruby.RADIOSITY.room.16p.16m.stats:
> L1D_cache_access_mode_type_SupervisorMode:   208131190    94.4257%
> RADIOSITY/ruby.RADIOSITY.room.16p.16m.stats:
> L1I_cache_access_mode_type_SupervisorMode:   3574778    28.1978%
> RADIOSITY/ruby.RADIOSITY.room.16p.16m.stats:
> L2_cache_access_mode_type_SupervisorMode:   99881703    97.481%
> RADIX/ruby.RADIX.1M.16p.16m.stats:
> L1D_cache_access_mode_type_SupervisorMode:   8949117    63.3698%
> RADIX/ruby.RADIX.1M.16p.16m.stats:
> L1I_cache_access_mode_type_SupervisorMode:   711276    93.5587%
> RADIX/ruby.RADIX.1M.16p.16m.stats:
> L2_cache_access_mode_type_SupervisorMode:   4191291    67.1359%
> RAYTRACE/ruby.RAYTRACE.car.16p.16m.stats:
> L1D_cache_access_mode_type_SupervisorMode:   5046186    42.2043%
> RAYTRACE/ruby.RAYTRACE.car.16p.16m.stats:
> L1I_cache_access_mode_type_SupervisorMode:   791814    8.57802%
> RAYTRACE/ruby.RAYTRACE.car.16p.16m.stats:
> L2_cache_access_mode_type_SupervisorMode:   2442022    31.3737%
> VOLREND/ruby.VOLREND.head.16p.16m.stats:
> L1D_cache_access_mode_type_SupervisorMode:   7130501    9.01353%
> VOLREND/ruby.VOLREND.head.16p.16m.stats:
> L1I_cache_access_mode_type_SupervisorMode:   5619630    37.7412%
> VOLREND/ruby.VOLREND.head.16p.16m.stats:
> L2_cache_access_mode_type_SupervisorMode:   3889090    71.005%
> WATER-SPATIAL/ruby.WATER-SPATIAL.512.16p.16m.stats:
> L1D_cache_access_mode_type_SupervisorMode:   10752580    53.7889%
> WATER-SPATIAL/ruby.WATER-SPATIAL.512.16p.16m.stats:
> L1I_cache_access_mode_type_SupervisorMode:   1149841    30.0857%
> WATER-SPATIAL/ruby.WATER-SPATIAL.512.16p.16m.stats:
> L2_cache_access_mode_type_SupervisorMode:   5518409    91.074%
>
>
> Then I turned on PROFILE_HOT_LINES flag, and got the following stat for FFT.
>
> Hot Data Blocks
> ---------------
>
> Total_entries_block_address: 75145
> Total_data_misses_block_address: 2474143
> total | load store atomic | user supervisor | sharing | touched-by
> block_address | 7.99368 % [0x300c0c0, line 0x300c0c0] 197775 | 197648
> 79 48 | 0 197775 | 0 | 16
> block_address | 7.98301 % [0x1b8920c0, line 0x1b8920c0] 197511 |
> 197349 115 47 | 0 197511 | 0 | 16
> block_address | 7.90132 % [0x1b0620c0, line 0x1b0620c0] 195490 |
> 195252 126 112 | 0 195490 | 0 | 16
> block_address | 7.82303 % [0x1b3c40c0, line 0x1b3c40c0] 193553 |
> 193443 73 37 | 0 193553 | 0 | 16
> block_address | 5.06996 % [0x1b0780c0, line 0x1b0780c0] 125438 |
> 125354 58 26 | 0 125438 | 0 | 16
> block_address | 4.83691 % [0xb6e080, line 0xb6e080] 119672 | 119672 0
> 0 | 0 119672 | 0 | 16
> ....
>
> Hot Instructions
> ----------------
>
> Total_entries_pc_address: 3889
> Total_data_misses_pc_address: 2474143
> total | load store atomic | user supervisor | sharing | touched-by
> pc_address | 30.4843 % [0x1055454, line 0x1055440] 754224 | 754224 0 0
> | 0 754224 | 0 | 16
> pc_address | 20.6657 % [0x1055458, line 0x1055440] 511298 | 511298 0 0
> | 0 511298 | 0 | 16
> pc_address | 19.1072 % [0x10554b4, line 0x1055480] 472740 | 472740 0 0
> | 0 472740 | 0 | 16
> pc_address | 5.30531 % [0x1053f50, line 0x1053f40] 131261 | 131261 0 0
> | 0 131261 | 0 | 16
> ......
>
>
> It seems like just a few instructions are executed on a few data
> blocks by the OS, making the statistics for each benchmark program
> almost meaningless. Any idea on what's happening inside the OS?
>
>
> Thanks,
> Ikhwan
> _______________________________________________
> Gems-users mailing list
> Gems-users@xxxxxxxxxxx
> https://lists.cs.wisc.edu/mailman/listinfo/gems-users
> Use Google to search the GEMS Users mailing list by adding "site:https://lists.cs.wisc.edu/archive/gems-users/" to your search.
>
>
_______________________________________________
Gems-users mailing list
Gems-users@xxxxxxxxxxx
https://lists.cs.wisc.edu/mailman/listinfo/gems-users
Use Google to search the GEMS Users mailing list by adding "site:https://lists.cs.wisc.edu/archive/gems-users/" to your search.

-----------------------------
David W Nellans
dnellans@xxxxxxxxxxx

_______________________________________________
Gems-users mailing list
Gems-users@xxxxxxxxxxx
https://lists.cs.wisc.edu/mailman/listinfo/gems-users
Use Google to search the GEMS Users mailing list by adding "site:https://lists.cs.wisc.edu/archive/gems-users/" to your search.



[← Prev in Thread] Current Thread [Next in Thread→]